code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = { "nvidia/segformer-b0-finetuned-ade-512-512": ( "https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class __magic_name__ ( __lowerCAmelCase): A: Any = "segformer" def __init__( self : Union[str, Any] , lowerCamelCase__ : int=3 , lowerCamelCase__ : str=4 , lowerCamelCase__ : List[Any]=[2, 2, 2, 2] , lowerCamelCase__ : int=[8, 4, 2, 1] , lowerCamelCase__ : Optional[int]=[32, 64, 160, 256] , lowerCamelCase__ : Any=[7, 3, 3, 3] , lowerCamelCase__ : str=[4, 2, 2, 2] , lowerCamelCase__ : Tuple=[1, 2, 5, 8] , lowerCamelCase__ : List[str]=[4, 4, 4, 4] , lowerCamelCase__ : Union[str, Any]="gelu" , lowerCamelCase__ : List[str]=0.0 , lowerCamelCase__ : Any=0.0 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[str]=0.02 , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : str=1E-6 , lowerCamelCase__ : Union[str, Any]=256 , lowerCamelCase__ : Tuple=255 , **lowerCamelCase__ : Optional[int] , ) -> Any: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( '''Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be''' ''' removed, as the behaviour will default to that of reshape_last_stage = True.''' , lowerCamelCase__ , ) UpperCamelCase__ : str = num_channels UpperCamelCase__ : Any = num_encoder_blocks UpperCamelCase__ : Optional[int] = depths UpperCamelCase__ : Optional[Any] = sr_ratios UpperCamelCase__ : Dict = hidden_sizes UpperCamelCase__ : Union[str, Any] = patch_sizes UpperCamelCase__ : Optional[Any] = strides UpperCamelCase__ : Union[str, Any] = mlp_ratios UpperCamelCase__ : str = num_attention_heads UpperCamelCase__ : Optional[Any] = hidden_act UpperCamelCase__ : List[Any] = hidden_dropout_prob UpperCamelCase__ : Dict = attention_probs_dropout_prob UpperCamelCase__ : str = classifier_dropout_prob UpperCamelCase__ : List[Any] = initializer_range UpperCamelCase__ : List[str] = drop_path_rate UpperCamelCase__ : str = layer_norm_eps UpperCamelCase__ : Any = decoder_hidden_size UpperCamelCase__ : List[str] = kwargs.get('''reshape_last_stage''' , lowerCamelCase__ ) UpperCamelCase__ : int = semantic_loss_ignore_index class __magic_name__ ( __lowerCAmelCase): A: Optional[Any] = version.parse("1.11") @property def UpperCAmelCase__ ( self : Any ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def UpperCAmelCase__ ( self : Union[str, Any] ) -> float: '''simple docstring''' return 1E-4 @property def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' return 12
146
from collections import deque from math import floor from random import random from time import time class __magic_name__ : def __init__( self : Optional[int] ) -> str: '''simple docstring''' UpperCamelCase__ : str = {} def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[int]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCamelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: UpperCamelCase__ : List[Any] = [[w, v]] if not self.graph.get(lowerCamelCase__ ): UpperCamelCase__ : Any = [] def UpperCAmelCase__ ( self : Optional[int] ) -> Any: '''simple docstring''' return list(self.graph ) def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Tuple ) -> Optional[int]: '''simple docstring''' if self.graph.get(lowerCamelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : int=-2 , lowerCamelCase__ : int=-1 ) -> List[Any]: '''simple docstring''' if s == d: return [] UpperCamelCase__ : List[str] = [] UpperCamelCase__ : Dict = [] if s == -2: UpperCamelCase__ : Optional[Any] = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Union[str, Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCamelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : Optional[int] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : int = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return visited def UpperCAmelCase__ ( self : str , lowerCamelCase__ : Optional[int]=-1 ) -> Optional[Any]: '''simple docstring''' if c == -1: UpperCamelCase__ : int = floor(random() * 10000 ) + 10 for i in range(lowerCamelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): UpperCamelCase__ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCamelCase__ , lowerCamelCase__ , 1 ) def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : Tuple=-2 ) -> str: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = deque() UpperCamelCase__ : Optional[Any] = [] if s == -2: UpperCamelCase__ : Optional[int] = list(self.graph )[0] d.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) while d: UpperCamelCase__ : List[str] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def UpperCAmelCase__ ( self : int , lowerCamelCase__ : List[Any] ) -> int: '''simple docstring''' UpperCamelCase__ : List[str] = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : List[str] ) -> int: '''simple docstring''' return len(self.graph[u] ) def UpperCAmelCase__ ( self : Dict , lowerCamelCase__ : List[str]=-2 ) -> Dict: '''simple docstring''' UpperCamelCase__ : int = [] UpperCamelCase__ : Optional[int] = [] if s == -2: UpperCamelCase__ : Dict = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : Optional[Any] = s UpperCamelCase__ : Dict = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Optional[int] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : Tuple = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : List[Any] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return sorted_nodes def UpperCAmelCase__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : Optional[int] = [] UpperCamelCase__ : int = [] UpperCamelCase__ : List[Any] = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : Dict = -2 UpperCamelCase__ : int = [] UpperCamelCase__ : Tuple = s UpperCamelCase__ : str = False UpperCamelCase__ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Dict = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): UpperCamelCase__ : Union[str, Any] = len(lowerCamelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() UpperCamelCase__ : Optional[int] = True if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : List[Any] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : Optional[Any] = False indirect_parents.append(lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = s UpperCamelCase__ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return list(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' UpperCamelCase__ : List[Any] = [] UpperCamelCase__ : Any = [] UpperCamelCase__ : Tuple = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : int = -2 UpperCamelCase__ : Optional[int] = [] UpperCamelCase__ : Tuple = s UpperCamelCase__ : List[str] = False UpperCamelCase__ : Tuple = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Any = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): UpperCamelCase__ : List[str] = len(lowerCamelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : int = node[1] break # check if all the children are visited if s == ss: stack.pop() UpperCamelCase__ : List[str] = True if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : Optional[Any] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : List[str] = False indirect_parents.append(lowerCamelCase__ ) UpperCamelCase__ : Tuple = s UpperCamelCase__ : List[Any] = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return False def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Union[str, Any]=-2 , lowerCamelCase__ : Union[str, Any]=-1 ) -> Any: '''simple docstring''' UpperCamelCase__ : Optional[int] = time() self.dfs(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : int = time() return end - begin def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : int=-2 ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ : List[str] = time() self.bfs(lowerCamelCase__ ) UpperCamelCase__ : Optional[Any] = time() return end - begin class __magic_name__ : def __init__( self : Optional[Any] ) -> Any: '''simple docstring''' UpperCamelCase__ : Dict = {} def UpperCAmelCase__ ( self : int , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Tuple=1 ) -> Dict: '''simple docstring''' if self.graph.get(lowerCamelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist UpperCamelCase__ : Union[str, Any] = [[w, v]] # add the other way if self.graph.get(lowerCamelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist UpperCamelCase__ : int = [[w, u]] def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) -> Tuple: '''simple docstring''' if self.graph.get(lowerCamelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCamelCase__ ) # the other way round if self.graph.get(lowerCamelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCamelCase__ ) def UpperCAmelCase__ ( self : int , lowerCamelCase__ : Tuple=-2 , lowerCamelCase__ : Tuple=-1 ) -> str: '''simple docstring''' if s == d: return [] UpperCamelCase__ : List[str] = [] UpperCamelCase__ : Tuple = [] if s == -2: UpperCamelCase__ : str = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : int = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : int = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCamelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : Any = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : Optional[Any] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : List[str] = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return visited def UpperCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[int]=-1 ) -> Optional[Any]: '''simple docstring''' if c == -1: UpperCamelCase__ : List[Any] = floor(random() * 10000 ) + 10 for i in range(lowerCamelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): UpperCamelCase__ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCamelCase__ , lowerCamelCase__ , 1 ) def UpperCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : int=-2 ) -> Tuple: '''simple docstring''' UpperCamelCase__ : List[Any] = deque() UpperCamelCase__ : int = [] if s == -2: UpperCamelCase__ : Dict = list(self.graph )[0] d.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) while d: UpperCamelCase__ : List[str] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def UpperCAmelCase__ ( self : int , lowerCamelCase__ : str ) -> List[Any]: '''simple docstring''' return len(self.graph[u] ) def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' UpperCamelCase__ : Optional[Any] = [] UpperCamelCase__ : Tuple = [] UpperCamelCase__ : str = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : Dict = -2 UpperCamelCase__ : Optional[Any] = [] UpperCamelCase__ : Optional[int] = s UpperCamelCase__ : int = False UpperCamelCase__ : str = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Tuple = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): UpperCamelCase__ : Optional[int] = len(lowerCamelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() UpperCamelCase__ : Optional[Any] = True if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : List[str] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : Optional[Any] = False indirect_parents.append(lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = s UpperCamelCase__ : Dict = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return list(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Any ) -> str: '''simple docstring''' UpperCamelCase__ : int = [] UpperCamelCase__ : str = [] UpperCamelCase__ : Optional[int] = list(self.graph )[0] stack.append(lowerCamelCase__ ) visited.append(lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = -2 UpperCamelCase__ : Union[str, Any] = [] UpperCamelCase__ : Optional[int] = s UpperCamelCase__ : str = False UpperCamelCase__ : Any = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: UpperCamelCase__ : Optional[int] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): UpperCamelCase__ : Optional[Any] = len(lowerCamelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) UpperCamelCase__ : int = node[1] break # check if all the children are visited if s == ss: stack.pop() UpperCamelCase__ : Optional[Any] = True if len(lowerCamelCase__ ) != 0: UpperCamelCase__ : Optional[int] = stack[len(lowerCamelCase__ ) - 1] else: UpperCamelCase__ : Tuple = False indirect_parents.append(lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = s UpperCamelCase__ : Dict = ss # check if se have reached the starting point if len(lowerCamelCase__ ) == 0: return False def UpperCAmelCase__ ( self : Dict ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : Any=-2 , lowerCamelCase__ : str=-1 ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : List[str] = time() self.dfs(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Dict = time() return end - begin def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : str=-2 ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : List[str] = time() self.bfs(lowerCamelCase__ ) UpperCamelCase__ : Any = time() return end - begin
146
1
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class lowercase__ : '''simple docstring''' a = BlenderbotConfig a = {} a = "gelu" def __init__( self, __magic_name__, __magic_name__=13, __magic_name__=7, __magic_name__=True, __magic_name__=False, __magic_name__=99, __magic_name__=32, __magic_name__=2, __magic_name__=4, __magic_name__=37, __magic_name__=0.1, __magic_name__=0.1, __magic_name__=20, __magic_name__=2, __magic_name__=1, __magic_name__=0, ) -> Any: """simple docstring""" UpperCamelCase__ : int = parent UpperCamelCase__ : int = batch_size UpperCamelCase__ : Tuple = seq_length UpperCamelCase__ : Optional[int] = is_training UpperCamelCase__ : Optional[Any] = use_labels UpperCamelCase__ : Any = vocab_size UpperCamelCase__ : List[Any] = hidden_size UpperCamelCase__ : Optional[Any] = num_hidden_layers UpperCamelCase__ : List[Any] = num_attention_heads UpperCamelCase__ : Optional[int] = intermediate_size UpperCamelCase__ : Union[str, Any] = hidden_dropout_prob UpperCamelCase__ : Optional[Any] = attention_probs_dropout_prob UpperCamelCase__ : Any = max_position_embeddings UpperCamelCase__ : Tuple = eos_token_id UpperCamelCase__ : Union[str, Any] = pad_token_id UpperCamelCase__ : Optional[Any] = bos_token_id def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ) UpperCamelCase__ : List[str] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ), 1 ) UpperCamelCase__ : Tuple = tf.concat([input_ids, eos_tensor], axis=1 ) UpperCamelCase__ : Any = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCamelCase__ : Union[str, Any] = self.config_cls( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_ids=[2], bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.pad_token_id, **self.config_updates, ) UpperCamelCase__ : List[Any] = prepare_blenderbot_inputs_dict(__magic_name__, __magic_name__, __magic_name__ ) return config, inputs_dict def UpperCamelCase__ ( self, __magic_name__, __magic_name__ ) -> Any: """simple docstring""" UpperCamelCase__ : int = TFBlenderbotModel(config=__magic_name__ ).get_decoder() UpperCamelCase__ : Union[str, Any] = inputs_dict['''input_ids'''] UpperCamelCase__ : Any = input_ids[:1, :] UpperCamelCase__ : int = inputs_dict['''attention_mask'''][:1, :] UpperCamelCase__ : str = inputs_dict['''head_mask'''] UpperCamelCase__ : Optional[int] = 1 # first forward pass UpperCamelCase__ : Dict = model(__magic_name__, attention_mask=__magic_name__, head_mask=__magic_name__, use_cache=__magic_name__ ) UpperCamelCase__ : Any = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCamelCase__ : List[Any] = ids_tensor((self.batch_size, 3), config.vocab_size ) UpperCamelCase__ : int = tf.cast(ids_tensor((self.batch_size, 3), 2 ), tf.inta ) # append to next input_ids and UpperCamelCase__ : Optional[int] = tf.concat([input_ids, next_tokens], axis=-1 ) UpperCamelCase__ : int = tf.concat([attention_mask, next_attn_mask], axis=-1 ) UpperCamelCase__ : Optional[int] = model(__magic_name__, attention_mask=__magic_name__ )[0] UpperCamelCase__ : str = model(__magic_name__, attention_mask=__magic_name__, past_key_values=__magic_name__ )[0] self.parent.assertEqual(next_tokens.shape[1], output_from_past.shape[1] ) # select random slice UpperCamelCase__ : str = int(ids_tensor((1,), output_from_past.shape[-1] ) ) UpperCamelCase__ : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx] UpperCamelCase__ : Any = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__magic_name__, __magic_name__, rtol=1E-3 ) def lowerCAmelCase_ ( __UpperCAmelCase: Optional[Any] , __UpperCAmelCase: int , __UpperCAmelCase: List[Any] , __UpperCAmelCase: int=None , __UpperCAmelCase: Union[str, Any]=None , __UpperCAmelCase: Union[str, Any]=None , __UpperCAmelCase: Optional[Any]=None , __UpperCAmelCase: str=None , ) -> Dict: if attention_mask is None: UpperCamelCase__ : Dict = tf.cast(tf.math.not_equal(__UpperCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCamelCase__ : Any = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCamelCase__ : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCamelCase__ : str = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCamelCase__ : Optional[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowercase__ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' a = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () a = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () a = ( { "conversational": TFBlenderbotForConditionalGeneration, "feature-extraction": TFBlenderbotModel, "summarization": TFBlenderbotForConditionalGeneration, "text2text-generation": TFBlenderbotForConditionalGeneration, "translation": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) a = True a = False a = False def UpperCamelCase__ ( self ) -> Any: """simple docstring""" UpperCamelCase__ : int = TFBlenderbotModelTester(self ) UpperCamelCase__ : int = ConfigTester(self, config_class=__magic_name__ ) def UpperCamelCase__ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) -> Dict: """simple docstring""" UpperCamelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__magic_name__ ) @require_tokenizers @require_tf class lowercase__ ( unittest.TestCase ): '''simple docstring''' a = ["My friends are cool but they eat too many carbs."] a = "facebook/blenderbot-400M-distill" @cached_property def UpperCamelCase__ ( self ) -> List[Any]: """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCamelCase__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase__ : Union[str, Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : Union[str, Any] = self.tokenizer(self.src_text, return_tensors='''tf''' ) UpperCamelCase__ : Optional[Any] = self.model.generate( model_inputs.input_ids, ) UpperCamelCase__ : Optional[int] = self.tokenizer.batch_decode(generated_ids.numpy(), skip_special_tokens=__magic_name__ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
352
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCAmelCase_ ( ) -> List[str]: UpperCamelCase__ : List[str] = { '''repo_name''': ['''test_repo1''', '''test_repo2''', '''test_repo3'''], '''path''': ['''test_1.py''', '''test_2.py''', '''unit_test.py'''], '''content''': ['''a ''' * 20, '''a ''' * 30, '''b ''' * 7], } UpperCamelCase__ : Dict = Dataset.from_dict(__UpperCAmelCase ) return dataset class lowercase__ ( __lowerCamelCase ): '''simple docstring''' def UpperCamelCase__ ( self ) -> Any: """simple docstring""" UpperCamelCase__ : List[Any] = get_dataset() UpperCamelCase__ : List[str] = make_duplicate_clusters(__magic_name__, 0.85 ) self.assertEqual(len(duplicate_clusters[0] ), 2 ) def UpperCamelCase__ ( self ) -> str: """simple docstring""" UpperCamelCase__ : List[Any] = get_dataset() UpperCamelCase__ ,UpperCamelCase__ : Dict = deduplicate_dataset(__magic_name__ ) self.assertEqual(len(__magic_name__ ), 2 ) print(__magic_name__ ) self.assertEqual(duplicate_clusters[0][0]['''copies'''], 2 ) self.assertEqual(duplicate_clusters[0][0]['''is_extreme'''], __magic_name__ )
247
0
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase__ ) class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = field(default='audio-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) SCREAMING_SNAKE_CASE = Features({'audio': Audio()} ) SCREAMING_SNAKE_CASE = Features({'labels': ClassLabel} ) SCREAMING_SNAKE_CASE = "audio" SCREAMING_SNAKE_CASE = "labels" def _a (self , _lowerCamelCase ): """simple docstring""" if self.label_column not in features: raise ValueError(F"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , _lowerCamelCase ): raise ValueError(F"""Column {self.label_column} is not a ClassLabel.""" ) UpperCAmelCase__ : List[Any] = copy.deepcopy(self ) UpperCAmelCase__ : Optional[Any] = self.label_schema.copy() UpperCAmelCase__ : List[Any] = features[self.label_column] UpperCAmelCase__ : List[Any] = label_schema return task_template @property def _a (self ): """simple docstring""" return { self.audio_column: "audio", self.label_column: "labels", }
171
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch _A = logging.get_logger(__name__) class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['pixel_values'] def __init__(self , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = PILImageResampling.BILINEAR , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = True , **_lowerCamelCase , ): """simple docstring""" super().__init__(**_lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = size if size is not None else {"""shortest_edge""": 224} UpperCAmelCase__ : List[Any] = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) UpperCAmelCase__ : str = crop_size if crop_size is not None else {"""height""": 256, """width""": 256} UpperCAmelCase__ : str = get_size_dict(_lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase__ : int = do_resize UpperCAmelCase__ : Any = size UpperCAmelCase__ : int = resample UpperCAmelCase__ : Union[str, Any] = do_rescale UpperCAmelCase__ : Union[str, Any] = rescale_factor UpperCAmelCase__ : str = do_center_crop UpperCAmelCase__ : Dict = crop_size UpperCAmelCase__ : List[str] = do_flip_channel_order def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = PIL.Image.BILINEAR , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : List[str] = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase__ : Union[str, Any] = get_resize_output_image_size(_lowerCamelCase , size=size["""shortest_edge"""] , default_to_square=_lowerCamelCase ) return resize(_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = get_size_dict(_lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowerCamelCase , size=(size["""height"""], size["""width"""]) , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" return rescale(_lowerCamelCase , scale=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" return flip_channel_order(_lowerCamelCase , data_format=_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = ChannelDimension.FIRST , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Any = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ : List[str] = resample if resample is not None else self.resample UpperCAmelCase__ : Tuple = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ : int = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) UpperCAmelCase__ : List[str] = size if size is not None else self.size UpperCAmelCase__ : Tuple = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) UpperCAmelCase__ : Optional[int] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ : str = get_size_dict(_lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase__ : List[str] = make_list_of_images(_lowerCamelCase ) if not valid_images(_lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase__ : Union[str, Any] = [to_numpy_array(_lowerCamelCase ) for image in images] if do_resize: UpperCAmelCase__ : Tuple = [self.resize(image=_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase ) for image in images] if do_center_crop: UpperCAmelCase__ : Optional[Any] = [self.center_crop(image=_lowerCamelCase , size=_lowerCamelCase ) for image in images] if do_rescale: UpperCAmelCase__ : List[Any] = [self.rescale(image=_lowerCamelCase , scale=_lowerCamelCase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: UpperCAmelCase__ : Any = [self.flip_channel_order(image=_lowerCamelCase ) for image in images] UpperCAmelCase__ : int = [to_channel_dimension_format(_lowerCamelCase , _lowerCamelCase ) for image in images] UpperCAmelCase__ : Optional[Any] = {"""pixel_values""": images} return BatchFeature(data=_lowerCamelCase , tensor_type=_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" UpperCAmelCase__ : List[str] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(_lowerCamelCase ): UpperCAmelCase__ : Optional[int] = target_sizes.numpy() UpperCAmelCase__ : Tuple = [] for idx in range(len(_lowerCamelCase ) ): UpperCAmelCase__ : Union[str, Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=_lowerCamelCase ) UpperCAmelCase__ : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowerCamelCase ) else: UpperCAmelCase__ : str = logits.argmax(dim=1 ) UpperCAmelCase__ : int = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
171
1
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _lowerCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a ) class __snake_case (_a ): def __init__( self : Dict , **_UpperCAmelCase : List[str] ) -> int: '''simple docstring''' super().__init__(**_UpperCAmelCase ) if self.framework == "tf": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) requires_backends(self , """vision""" ) self.check_model_type(_UpperCAmelCase ) def __call__( self : Optional[int] , _UpperCAmelCase : Union[str, "Image.Image", List[Dict[str, Any]]] , _UpperCAmelCase : Union[str, List[str]] = None , **_UpperCAmelCase : List[str] , ) -> str: '''simple docstring''' if "text_queries" in kwargs: _lowerCAmelCase : str = kwargs.pop("""text_queries""" ) if isinstance(_UpperCAmelCase , (str, Image.Image) ): _lowerCAmelCase : Union[str, Any] = {"""image""": image, """candidate_labels""": candidate_labels} else: _lowerCAmelCase : Optional[Any] = image _lowerCAmelCase : Any = super().__call__(_UpperCAmelCase , **_UpperCAmelCase ) return results def SCREAMING_SNAKE_CASE ( self : List[str] , **_UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _lowerCAmelCase : Optional[Any] = {} if "threshold" in kwargs: _lowerCAmelCase : Dict = kwargs["""threshold"""] if "top_k" in kwargs: _lowerCAmelCase : List[Any] = kwargs["""top_k"""] return {}, {}, postprocess_params def SCREAMING_SNAKE_CASE ( self : Tuple , _UpperCAmelCase : Optional[int] ) -> str: '''simple docstring''' _lowerCAmelCase : List[Any] = load_image(inputs["""image"""] ) _lowerCAmelCase : Dict = inputs["""candidate_labels"""] if isinstance(_UpperCAmelCase , _UpperCAmelCase ): _lowerCAmelCase : Optional[Any] = candidate_labels.split(""",""" ) _lowerCAmelCase : Any = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(_UpperCAmelCase ): _lowerCAmelCase : Dict = self.tokenizer(_UpperCAmelCase , return_tensors=self.framework ) _lowerCAmelCase : Any = self.image_processor(_UpperCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(_UpperCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def SCREAMING_SNAKE_CASE ( self : Optional[int] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: '''simple docstring''' _lowerCAmelCase : str = model_inputs.pop("""target_size""" ) _lowerCAmelCase : Dict = model_inputs.pop("""candidate_label""" ) _lowerCAmelCase : Optional[Any] = model_inputs.pop("""is_last""" ) _lowerCAmelCase : Any = self.model(**_UpperCAmelCase ) _lowerCAmelCase : str = {"""target_size""": target_size, """candidate_label""": candidate_label, """is_last""": is_last, **outputs} return model_outputs def SCREAMING_SNAKE_CASE ( self : Optional[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Optional[Any]=0.1 , _UpperCAmelCase : Union[str, Any]=None ) -> List[str]: '''simple docstring''' _lowerCAmelCase : str = [] for model_output in model_outputs: _lowerCAmelCase : Tuple = model_output["""candidate_label"""] _lowerCAmelCase : Union[str, Any] = BaseModelOutput(_UpperCAmelCase ) _lowerCAmelCase : int = self.image_processor.post_process_object_detection( outputs=_UpperCAmelCase , threshold=_UpperCAmelCase , target_sizes=model_output["""target_size"""] )[0] for index in outputs["scores"].nonzero(): _lowerCAmelCase : str = outputs["""scores"""][index].item() _lowerCAmelCase : str = self._get_bounding_box(outputs["""boxes"""][index][0] ) _lowerCAmelCase : List[Any] = {"""score""": score, """label""": label, """box""": box} results.append(_UpperCAmelCase ) _lowerCAmelCase : Union[str, Any] = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x["score"] , reverse=_UpperCAmelCase ) if top_k: _lowerCAmelCase : int = results[:top_k] return results def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , _UpperCAmelCase : "torch.Tensor" ) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError("""The ZeroShotObjectDetectionPipeline is only available in PyTorch.""" ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : str = box.int().tolist() _lowerCAmelCase : str = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
159
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Any = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _lowerCamelCase : Tuple = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.encoder.norm.weight", "encoder.layernorm.weight"), ("transformer.encoder.norm.bias", "encoder.layernorm.bias"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) def _UpperCAmelCase (UpperCamelCase_ : Dict , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Optional[Any] ): '''simple docstring''' _lowerCAmelCase : str = state_dict.pop(UpperCamelCase_ ) _lowerCAmelCase : Tuple = val def _UpperCAmelCase (UpperCamelCase_ : Union[str, Any] ): '''simple docstring''' _lowerCAmelCase : Optional[int] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _lowerCAmelCase : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) _lowerCAmelCase : Optional[Any] = value else: _lowerCAmelCase : Union[str, Any] = value return new_state_dict def _UpperCAmelCase (UpperCamelCase_ : Optional[int] ): '''simple docstring''' _lowerCAmelCase : Tuple = """""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowerCAmelCase : Dict = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) _lowerCAmelCase : str = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict _lowerCAmelCase : str = in_proj_weight[:256, :] _lowerCAmelCase : Tuple = in_proj_bias[:256] _lowerCAmelCase : Tuple = in_proj_weight[256:512, :] _lowerCAmelCase : Dict = in_proj_bias[256:512] _lowerCAmelCase : Tuple = in_proj_weight[-256:, :] _lowerCAmelCase : List[str] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowerCAmelCase : Optional[int] = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) _lowerCAmelCase : Any = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict _lowerCAmelCase : Optional[Any] = in_proj_weight[:256, :] _lowerCAmelCase : Dict = in_proj_bias[:256] _lowerCAmelCase : Optional[Any] = in_proj_weight[256:512, :] _lowerCAmelCase : Optional[Any] = in_proj_bias[256:512] _lowerCAmelCase : str = in_proj_weight[-256:, :] _lowerCAmelCase : Any = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowerCAmelCase : Union[str, Any] = state_dict.pop( F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight" ) _lowerCAmelCase : str = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias" ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowerCAmelCase : Tuple = in_proj_weight_cross_attn[:256, :] _lowerCAmelCase : List[str] = in_proj_bias_cross_attn[:256] _lowerCAmelCase : Dict = in_proj_weight_cross_attn[256:512, :] _lowerCAmelCase : Any = in_proj_bias_cross_attn[256:512] _lowerCAmelCase : Tuple = in_proj_weight_cross_attn[-256:, :] _lowerCAmelCase : List[Any] = in_proj_bias_cross_attn[-256:] def _UpperCAmelCase (UpperCamelCase_ : List[Any] , UpperCamelCase_ : Any ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = image.size _lowerCAmelCase : List[str] = max(UpperCamelCase_ , UpperCamelCase_ ) _lowerCAmelCase : Union[str, Any] = 800 if """detection""" in checkpoint_url else 1000 _lowerCAmelCase : Optional[int] = target_max_size / current_max_size _lowerCAmelCase : Union[str, Any] = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _UpperCAmelCase (UpperCamelCase_ : Any ): '''simple docstring''' _lowerCAmelCase : Optional[int] = F.to_tensor(UpperCamelCase_ ) _lowerCAmelCase : int = F.normalize(UpperCamelCase_ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _UpperCAmelCase (UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Optional[int] ): '''simple docstring''' logger.info("""Converting model...""" ) # load original state dict _lowerCAmelCase : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase_ , map_location="""cpu""" ) # rename keys for src, dest in rename_keys: rename_key(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) _lowerCAmelCase : List[str] = rename_backbone_keys(UpperCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowerCAmelCase : int = """model.""" for key in state_dict.copy().keys(): if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): _lowerCAmelCase : Tuple = state_dict.pop(UpperCamelCase_ ) _lowerCAmelCase : str = val # create HuggingFace model and load state dict _lowerCAmelCase : List[Any] = TableTransformerConfig( backbone="""resnet18""" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _lowerCAmelCase : Any = 15 _lowerCAmelCase : Any = 2 _lowerCAmelCase : List[Any] = {0: """table""", 1: """table rotated"""} _lowerCAmelCase : Tuple = idalabel _lowerCAmelCase : Optional[Any] = {v: k for k, v in idalabel.items()} else: _lowerCAmelCase : Optional[Any] = 125 _lowerCAmelCase : int = 6 _lowerCAmelCase : List[str] = { 0: """table""", 1: """table column""", 2: """table row""", 3: """table column header""", 4: """table projected row header""", 5: """table spanning cell""", } _lowerCAmelCase : int = idalabel _lowerCAmelCase : Dict = {v: k for k, v in idalabel.items()} _lowerCAmelCase : Any = DetrImageProcessor( format="""coco_detection""" , max_size=800 if """detection""" in checkpoint_url else 1000 ) _lowerCAmelCase : int = TableTransformerForObjectDetection(UpperCamelCase_ ) model.load_state_dict(UpperCamelCase_ ) model.eval() # verify our conversion _lowerCAmelCase : Optional[Any] = """example_pdf.png""" if """detection""" in checkpoint_url else """example_table.png""" _lowerCAmelCase : Union[str, Any] = hf_hub_download(repo_id="""nielsr/example-pdf""" , repo_type="""dataset""" , filename=UpperCamelCase_ ) _lowerCAmelCase : Union[str, Any] = Image.open(UpperCamelCase_ ).convert("""RGB""" ) _lowerCAmelCase : Optional[Any] = normalize(resize(UpperCamelCase_ , UpperCamelCase_ ) ).unsqueeze(0 ) _lowerCAmelCase : Optional[Any] = model(UpperCamelCase_ ) if "detection" in checkpoint_url: _lowerCAmelCase : Dict = (1, 15, 3) _lowerCAmelCase : Optional[Any] = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) _lowerCAmelCase : Union[str, Any] = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: _lowerCAmelCase : Optional[Any] = (1, 125, 7) _lowerCAmelCase : Tuple = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) _lowerCAmelCase : Dict = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , UpperCamelCase_ , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , UpperCamelCase_ , atol=1E-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(UpperCamelCase_ ).mkdir(exist_ok=UpperCamelCase_ ) model.save_pretrained(UpperCamelCase_ ) image_processor.save_pretrained(UpperCamelCase_ ) if push_to_hub: # Push model to HF hub logger.info("""Pushing model to the hub...""" ) _lowerCAmelCase : List[str] = ( """microsoft/table-transformer-detection""" if """detection""" in checkpoint_url else """microsoft/table-transformer-structure-recognition""" ) model.push_to_hub(UpperCamelCase_ ) image_processor.push_to_hub(UpperCamelCase_ ) if __name__ == "__main__": _lowerCamelCase : Dict = argparse.ArgumentParser() parser.add_argument( "--checkpoint_url", default="https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", type=str, choices=[ "https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", "https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth", ], help="URL of the Table Transformer checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _lowerCamelCase : List[str] = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
159
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ : List[str] = logging.get_logger(__name__) UpperCAmelCase_ : int = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Union[str, Any] = """blip_2_vision_model""" def __init__( self : Any , __lowerCamelCase : Any=1_408 , __lowerCamelCase : Any=6_144 , __lowerCamelCase : Any=39 , __lowerCamelCase : List[str]=16 , __lowerCamelCase : List[Any]=224 , __lowerCamelCase : List[str]=14 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : List[str]=0.00001 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : List[str]=1E-10 , __lowerCamelCase : int=True , **__lowerCamelCase : str , ): super().__init__(**__lowerCamelCase ) UpperCamelCase :Optional[Any] = hidden_size UpperCamelCase :List[str] = intermediate_size UpperCamelCase :Optional[Any] = num_hidden_layers UpperCamelCase :Union[str, Any] = num_attention_heads UpperCamelCase :List[str] = patch_size UpperCamelCase :str = image_size UpperCamelCase :Optional[int] = initializer_range UpperCamelCase :Optional[int] = attention_dropout UpperCamelCase :Union[str, Any] = layer_norm_eps UpperCamelCase :Optional[int] = hidden_act UpperCamelCase :List[str] = qkv_bias @classmethod def _A ( cls : Tuple , __lowerCamelCase : Union[str, os.PathLike] , **__lowerCamelCase : List[str] ): cls._set_token_in_kwargs(__lowerCamelCase ) UpperCamelCase , UpperCamelCase :List[Any] = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase :Union[str, Any] = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : int = """blip_2_qformer""" def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=30_522 , __lowerCamelCase : Any=768 , __lowerCamelCase : Union[str, Any]=12 , __lowerCamelCase : Optional[Any]=12 , __lowerCamelCase : List[str]=3_072 , __lowerCamelCase : Union[str, Any]="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=512 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Union[str, Any]=1E-12 , __lowerCamelCase : Any=0 , __lowerCamelCase : Tuple="absolute" , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : Optional[Any]=1_408 , **__lowerCamelCase : str , ): super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) UpperCamelCase :Tuple = vocab_size UpperCamelCase :str = hidden_size UpperCamelCase :Dict = num_hidden_layers UpperCamelCase :Tuple = num_attention_heads UpperCamelCase :Optional[Any] = hidden_act UpperCamelCase :Any = intermediate_size UpperCamelCase :Tuple = hidden_dropout_prob UpperCamelCase :List[str] = attention_probs_dropout_prob UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :Union[str, Any] = initializer_range UpperCamelCase :Tuple = layer_norm_eps UpperCamelCase :Tuple = position_embedding_type UpperCamelCase :Optional[Any] = cross_attention_frequency UpperCamelCase :Optional[Any] = encoder_hidden_size @classmethod def _A ( cls : Optional[Any] , __lowerCamelCase : Union[str, os.PathLike] , **__lowerCamelCase : int ): cls._set_token_in_kwargs(__lowerCamelCase ) UpperCamelCase , UpperCamelCase :str = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase :str = config_dict["""qformer_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : int = """blip-2""" snake_case__ : Optional[Any] = True def __init__( self : int , __lowerCamelCase : Dict=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : Dict=32 , **__lowerCamelCase : Dict ): super().__init__(**__lowerCamelCase ) if vision_config is None: UpperCamelCase :List[str] = {} logger.info("""vision_config is None. initializing the Blip2VisionConfig with default values.""" ) if qformer_config is None: UpperCamelCase :List[Any] = {} logger.info("""qformer_config is None. Initializing the Blip2QFormerConfig with default values.""" ) if text_config is None: UpperCamelCase :int = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) UpperCamelCase :Tuple = BlipaVisionConfig(**__lowerCamelCase ) UpperCamelCase :str = BlipaQFormerConfig(**__lowerCamelCase ) UpperCamelCase :Any = text_config["""model_type"""] if """model_type""" in text_config else """opt""" UpperCamelCase :Union[str, Any] = CONFIG_MAPPING[text_model_type](**__lowerCamelCase ) UpperCamelCase :List[str] = self.text_config.tie_word_embeddings UpperCamelCase :Union[str, Any] = self.text_config.is_encoder_decoder UpperCamelCase :Tuple = num_query_tokens UpperCamelCase :Any = self.vision_config.hidden_size UpperCamelCase :int = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES UpperCamelCase :Union[str, Any] = 1.0 UpperCamelCase :Union[str, Any] = 0.02 @classmethod def _A ( cls : Tuple , __lowerCamelCase : BlipaVisionConfig , __lowerCamelCase : BlipaQFormerConfig , __lowerCamelCase : PretrainedConfig , **__lowerCamelCase : List[Any] , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__lowerCamelCase , ) def _A ( self : Optional[int] ): UpperCamelCase :Tuple = copy.deepcopy(self.__dict__ ) UpperCamelCase :Dict = self.vision_config.to_dict() UpperCamelCase :Optional[int] = self.qformer_config.to_dict() UpperCamelCase :Any = self.text_config.to_dict() UpperCamelCase :Union[str, Any] = self.__class__.model_type return output
38
'''simple docstring''' import numpy as np import pandas as pd from sklearn.preprocessing import Normalizer from sklearn.svm import SVR from statsmodels.tsa.statespace.sarimax import SARIMAX def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )-> float: """simple docstring""" __A = np.array([[1, item, train_mtch[i]] for i, item in enumerate(UpperCAmelCase )] ) __A = np.array(UpperCAmelCase ) __A = np.dot(np.dot(np.linalg.inv(np.dot(x.transpose() , UpperCAmelCase ) ) , x.transpose() ) , UpperCAmelCase ) return abs(beta[0] + test_dt[0] * beta[1] + test_mtch[0] + beta[2] ) def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )-> float: """simple docstring""" __A = (1, 2, 1) __A = (1, 1, 0, 7) __A = SARIMAX( UpperCAmelCase , exog=UpperCAmelCase , order=UpperCAmelCase , seasonal_order=UpperCAmelCase ) __A = model.fit(disp=UpperCAmelCase , maxiter=6_0_0 , method='nm' ) __A = model_fit.predict(1 , len(UpperCAmelCase ) , exog=[test_match] ) return result[0] def snake_case ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )-> float: """simple docstring""" __A = SVR(kernel='rbf' , C=1 , gamma=0.1 , epsilon=0.1 ) regressor.fit(UpperCAmelCase , UpperCAmelCase ) __A = regressor.predict(UpperCAmelCase ) return y_pred[0] def snake_case ( UpperCAmelCase )-> float: """simple docstring""" train_user.sort() __A = np.percentile(UpperCAmelCase , 2_5 ) __A = np.percentile(UpperCAmelCase , 7_5 ) __A = qa - qa __A = qa - (iqr * 0.1) return low_lim def snake_case ( UpperCAmelCase , UpperCAmelCase )-> bool: """simple docstring""" __A = 0 __A = 0 for i in list_vote: if i > actual_result: __A = not_safe + 1 else: if abs(abs(UpperCAmelCase ) - abs(UpperCAmelCase ) ) <= 0.1: safe += 1 else: not_safe += 1 return safe > not_safe if __name__ == "__main__": # data_input_df = pd.read_csv("ex_data.csv", header=None) a__ : List[str] = [[1_8_2_3_1, 0.0, 1], [2_2_6_2_1, 1.0, 2], [1_5_6_7_5, 0.0, 3], [2_3_5_8_3, 1.0, 4]] a__ : Optional[int] = pd.DataFrame( data_input, columns=["total_user", "total_even", "days"] ) a__ : List[Any] = Normalizer().fit_transform(data_input_df.values) # split data a__ : Dict = normalize_df[:, 2].tolist() a__ : Optional[int] = normalize_df[:, 0].tolist() a__ : str = normalize_df[:, 1].tolist() # for svr (input variable = total date and total match) a__ : Tuple = normalize_df[:, [1, 2]].tolist() a__ : Dict = x[: len(x) - 1] a__ : Any = x[len(x) - 1 :] # for linear regression & sarimax a__ : Tuple = total_date[: len(total_date) - 1] a__ : List[Any] = total_user[: len(total_user) - 1] a__ : List[Any] = total_match[: len(total_match) - 1] a__ : List[str] = total_date[len(total_date) - 1 :] a__ : List[str] = total_user[len(total_user) - 1 :] a__ : Tuple = total_match[len(total_match) - 1 :] # voting system with forecasting a__ : Optional[Any] = [ linear_regression_prediction( trn_date, trn_user, trn_match, tst_date, tst_match ), sarimax_predictor(trn_user, trn_match, tst_match), support_vector_regressor(x_train, x_test, trn_user), ] # check the safety of today's data a__ : List[str] = "" if data_safety_checker(res_vote, tst_user) else "not " print("Today's data is {not_str}safe.")
161
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor _lowerCamelCase : List[Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : Union[str, Any] , *UpperCamelCase__ : Any , **UpperCamelCase__ : Union[str, Any] ): """simple docstring""" warnings.warn( 'The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PoolFormerImageProcessor instead.' , UpperCamelCase__ , ) super().__init__(*UpperCamelCase__ , **UpperCamelCase__ )
249
'''simple docstring''' def __lowerCamelCase ( A__ , A__ ) -> float: """simple docstring""" if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modulus' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
249
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: if digit_amount > 0: return round(number - int(UpperCamelCase ) , UpperCamelCase ) return number - int(UpperCamelCase ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
41
'''simple docstring''' import math def a__ ( lowercase : list, lowercase : int = 0, lowercase : int = 0 ) -> list: """simple docstring""" _UpperCamelCase = end or len(lowercase ) for i in range(lowercase, lowercase ): _UpperCamelCase = i _UpperCamelCase = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: _UpperCamelCase = array[temp_index - 1] temp_index -= 1 _UpperCamelCase = temp_index_value return array def a__ ( lowercase : list, lowercase : int, lowercase : int ) -> None: # Max Heap """simple docstring""" _UpperCamelCase = index _UpperCamelCase = 2 * index + 1 # Left Node _UpperCamelCase = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: _UpperCamelCase = left_index if right_index < heap_size and array[largest] < array[right_index]: _UpperCamelCase = right_index if largest != index: _UpperCamelCase , _UpperCamelCase = array[largest], array[index] heapify(lowercase, lowercase, lowercase ) def a__ ( lowercase : list ) -> list: """simple docstring""" _UpperCamelCase = len(lowercase ) for i in range(n // 2, -1, -1 ): heapify(lowercase, lowercase, lowercase ) for i in range(n - 1, 0, -1 ): _UpperCamelCase , _UpperCamelCase = array[0], array[i] heapify(lowercase, 0, lowercase ) return array def a__ ( lowercase : list, lowercase : int, lowercase : int, lowercase : int ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def a__ ( lowercase : list, lowercase : int, lowercase : int, lowercase : int ) -> int: """simple docstring""" _UpperCamelCase = low _UpperCamelCase = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i _UpperCamelCase , _UpperCamelCase = array[j], array[i] i += 1 def a__ ( lowercase : list ) -> list: """simple docstring""" if len(lowercase ) == 0: return array _UpperCamelCase = 2 * math.ceil(math.loga(len(lowercase ) ) ) _UpperCamelCase = 16 return intro_sort(lowercase, 0, len(lowercase ), lowercase, lowercase ) def a__ ( lowercase : list, lowercase : int, lowercase : int, lowercase : int, lowercase : int ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(lowercase ) max_depth -= 1 _UpperCamelCase = median_of_a(lowercase, lowercase, start + ((end - start) // 2) + 1, end - 1 ) _UpperCamelCase = partition(lowercase, lowercase, lowercase, lowercase ) intro_sort(lowercase, lowercase, lowercase, lowercase, lowercase ) _UpperCamelCase = p return insertion_sort(lowercase, lowercase, lowercase ) if __name__ == "__main__": import doctest doctest.testmod() lowercase__ : Any = input('Enter numbers separated by a comma : ').strip() lowercase__ : Any = [float(item) for item in user_input.split(',')] print(sort(unsorted))
324
0
'''simple docstring''' import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a : str = logging.get_logger() def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase = True ) -> Optional[Any]: print(f'Converting {name}...' ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": __lowerCAmelCase = timm.create_model("""levit_128s""" , pretrained=lowercase ) else: __lowerCAmelCase = timm.create_model("""levit_128""" , pretrained=lowercase ) if hidden_sizes == 192: __lowerCAmelCase = timm.create_model("""levit_192""" , pretrained=lowercase ) if hidden_sizes == 256: __lowerCAmelCase = timm.create_model("""levit_256""" , pretrained=lowercase ) if hidden_sizes == 384: __lowerCAmelCase = timm.create_model("""levit_384""" , pretrained=lowercase ) from_model.eval() __lowerCAmelCase = LevitForImageClassificationWithTeacher(lowercase ).eval() __lowerCAmelCase = OrderedDict() __lowerCAmelCase = from_model.state_dict() __lowerCAmelCase = list(from_model.state_dict().keys() ) __lowerCAmelCase = list(our_model.state_dict().keys() ) print(len(lowercase ) , len(lowercase ) ) for i in range(len(lowercase ) ): __lowerCAmelCase = weights[og_keys[i]] our_model.load_state_dict(lowercase ) __lowerCAmelCase = torch.randn((2, 3, 224, 224) ) __lowerCAmelCase = from_model(lowercase ) __lowerCAmelCase = our_model(lowercase ).logits assert torch.allclose(lowercase , lowercase ), "The model logits don't match the original one." __lowerCAmelCase = name print(lowercase ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) __lowerCAmelCase = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(f'Pushed {checkpoint_name}' ) def _lowerCAmelCase ( lowercase , lowercase = None , lowercase = True ) -> Any: __lowerCAmelCase = """imagenet-1k-id2label.json""" __lowerCAmelCase = 1000 __lowerCAmelCase = (1, num_labels) __lowerCAmelCase = """huggingface/label-files""" __lowerCAmelCase = num_labels __lowerCAmelCase = json.load(open(hf_hub_download(lowercase , lowercase , repo_type="""dataset""" ) , """r""" ) ) __lowerCAmelCase = {int(lowercase ): v for k, v in idalabel.items()} __lowerCAmelCase = idalabel __lowerCAmelCase = {v: k for k, v in idalabel.items()} __lowerCAmelCase = partial(lowercase , num_labels=lowercase , idalabel=lowercase , labelaid=lowercase ) __lowerCAmelCase = { """levit-128S""": 128, """levit-128""": 128, """levit-192""": 192, """levit-256""": 256, """levit-384""": 384, } __lowerCAmelCase = { """levit-128S""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-128""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-192""": ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-256""": ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-384""": ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , lowercase , names_to_config[model_name] , lowercase , lowercase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , lowercase , lowercase , lowercase , lowercase ) return config, expected_shape if __name__ == "__main__": _a : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default=None, type=str, help="""The name of the model you wish to convert, it must be one of the supported Levit* architecture,""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""levit-dump-folder/""", type=Path, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") parser.add_argument( """--no-push_to_hub""", dest="""push_to_hub""", action="""store_false""", help="""Do not push model and image processor to the hub""", ) _a : str = parser.parse_args() _a : Path = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
351
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings @maybe_allow_in_graph class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=0.0,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = "geglu",__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = True,__SCREAMING_SNAKE_CASE = "layer_norm",__SCREAMING_SNAKE_CASE = False,): '''simple docstring''' super().__init__() __lowerCAmelCase = only_cross_attention __lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == """ada_norm_zero""" __lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == """ada_norm""" if norm_type in ("ada_norm", "ada_norm_zero") and num_embeds_ada_norm is None: raise ValueError( f'`norm_type` is set to {norm_type}, but `num_embeds_ada_norm` is not defined. Please make sure to' f' define `num_embeds_ada_norm` if setting `norm_type` to {norm_type}.' ) # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn if self.use_ada_layer_norm: __lowerCAmelCase = AdaLayerNorm(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif self.use_ada_layer_norm_zero: __lowerCAmelCase = AdaLayerNormZero(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) else: __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = Attention( query_dim=__SCREAMING_SNAKE_CASE,heads=__SCREAMING_SNAKE_CASE,dim_head=__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,bias=__SCREAMING_SNAKE_CASE,cross_attention_dim=cross_attention_dim if only_cross_attention else None,upcast_attention=__SCREAMING_SNAKE_CASE,) # 2. Cross-Attn if cross_attention_dim is not None or double_self_attention: # We currently only use AdaLayerNormZero for self attention where there will only be one attention block. # I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during # the second cross attention block. __lowerCAmelCase = ( AdaLayerNorm(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm else nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) ) __lowerCAmelCase = Attention( query_dim=__SCREAMING_SNAKE_CASE,cross_attention_dim=cross_attention_dim if not double_self_attention else None,heads=__SCREAMING_SNAKE_CASE,dim_head=__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,bias=__SCREAMING_SNAKE_CASE,upcast_attention=__SCREAMING_SNAKE_CASE,) # is self-attn if encoder_hidden_states is none else: __lowerCAmelCase = None __lowerCAmelCase = None # 3. Feed-forward __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = FeedForward(__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,activation_fn=__SCREAMING_SNAKE_CASE,final_dropout=__SCREAMING_SNAKE_CASE ) # let chunk size default to None __lowerCAmelCase = None __lowerCAmelCase = 0 def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = chunk_size __lowerCAmelCase = dim def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,): '''simple docstring''' if self.use_ada_layer_norm: __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif self.use_ada_layer_norm_zero: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.norma( __SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,hidden_dtype=hidden_states.dtype ) else: __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = cross_attention_kwargs if cross_attention_kwargs is not None else {} __lowerCAmelCase = self.attna( __SCREAMING_SNAKE_CASE,encoder_hidden_states=encoder_hidden_states if self.only_cross_attention else None,attention_mask=__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE,) if self.use_ada_layer_norm_zero: __lowerCAmelCase = gate_msa.unsqueeze(1 ) * attn_output __lowerCAmelCase = attn_output + hidden_states # 2. Cross-Attention if self.attna is not None: __lowerCAmelCase = ( self.norma(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm else self.norma(__SCREAMING_SNAKE_CASE ) ) __lowerCAmelCase = self.attna( __SCREAMING_SNAKE_CASE,encoder_hidden_states=__SCREAMING_SNAKE_CASE,attention_mask=__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE,) __lowerCAmelCase = attn_output + hidden_states # 3. Feed-forward __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm_zero: __lowerCAmelCase = norm_hidden_states * (1 + scale_mlp[:, None]) + shift_mlp[:, None] if self._chunk_size is not None: # "feed_forward_chunk_size" can be used to save memory if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0: raise ValueError( f'`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`.' ) __lowerCAmelCase = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size __lowerCAmelCase = torch.cat( [self.ff(__SCREAMING_SNAKE_CASE ) for hid_slice in norm_hidden_states.chunk(__SCREAMING_SNAKE_CASE,dim=self._chunk_dim )],dim=self._chunk_dim,) else: __lowerCAmelCase = self.ff(__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm_zero: __lowerCAmelCase = gate_mlp.unsqueeze(1 ) * ff_output __lowerCAmelCase = ff_output + hidden_states return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = 4,__SCREAMING_SNAKE_CASE = 0.0,__SCREAMING_SNAKE_CASE = "geglu",__SCREAMING_SNAKE_CASE = False,): '''simple docstring''' super().__init__() __lowerCAmelCase = int(dim * mult ) __lowerCAmelCase = dim_out if dim_out is not None else dim if activation_fn == "gelu": __lowerCAmelCase = GELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if activation_fn == "gelu-approximate": __lowerCAmelCase = GELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,approximate="""tanh""" ) elif activation_fn == "geglu": __lowerCAmelCase = GEGLU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif activation_fn == "geglu-approximate": __lowerCAmelCase = ApproximateGELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.ModuleList([] ) # project in self.net.append(__SCREAMING_SNAKE_CASE ) # project dropout self.net.append(nn.Dropout(__SCREAMING_SNAKE_CASE ) ) # project out self.net.append(nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) ) # FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout if final_dropout: self.net.append(nn.Dropout(__SCREAMING_SNAKE_CASE ) ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' for module in self.net: __lowerCAmelCase = module(__SCREAMING_SNAKE_CASE ) return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = "none" ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = approximate def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if gate.device.type != "mps": return F.gelu(__SCREAMING_SNAKE_CASE,approximate=self.approximate ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ),approximate=self.approximate ).to(dtype=gate.dtype ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = self.gelu(__SCREAMING_SNAKE_CASE ) return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,dim_out * 2 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if gate.device.type != "mps": return F.gelu(__SCREAMING_SNAKE_CASE ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) ).to(dtype=gate.dtype ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ).chunk(2,dim=-1 ) return hidden_states * self.gelu(__SCREAMING_SNAKE_CASE ) class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ) return x * torch.sigmoid(1.702 * x ) class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Embedding(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.SiLU() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,embedding_dim * 2 ) __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.linear(self.silu(self.emb(__SCREAMING_SNAKE_CASE ) ) ) __lowerCAmelCase , __lowerCAmelCase = torch.chunk(__SCREAMING_SNAKE_CASE,2 ) __lowerCAmelCase = self.norm(__SCREAMING_SNAKE_CASE ) * (1 + scale) + shift return x class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = CombinedTimestepLabelEmbeddings(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.SiLU() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,6 * embedding_dim,bias=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE,eps=1e-6 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=None ): '''simple docstring''' __lowerCAmelCase = self.linear(self.silu(self.emb(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,hidden_dtype=__SCREAMING_SNAKE_CASE ) ) ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = emb.chunk(6,dim=1 ) __lowerCAmelCase = self.norm(__SCREAMING_SNAKE_CASE ) * (1 + scale_msa[:, None]) + shift_msa[:, None] return x, gate_msa, shift_mlp, scale_mlp, gate_mlp class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = 1e-5 ): '''simple docstring''' super().__init__() __lowerCAmelCase = num_groups __lowerCAmelCase = eps if act_fn is None: __lowerCAmelCase = None else: __lowerCAmelCase = get_activation(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,out_dim * 2 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if self.act: __lowerCAmelCase = self.act(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = self.linear(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = emb[:, :, None, None] __lowerCAmelCase , __lowerCAmelCase = emb.chunk(2,dim=1 ) __lowerCAmelCase = F.group_norm(__SCREAMING_SNAKE_CASE,self.num_groups,eps=self.eps ) __lowerCAmelCase = x * (1 + scale) + shift return x
46
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: SCREAMING_SNAKE_CASE_: Optional[Any] =None SCREAMING_SNAKE_CASE_: List[Any] =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Tuple ={'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} SCREAMING_SNAKE_CASE_: Union[str, Any] ={ 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } SCREAMING_SNAKE_CASE_: Optional[int] ={ 'facebook/mbart-large-en-ro': 10_24, 'facebook/mbart-large-cc25': 10_24, } # fmt: off SCREAMING_SNAKE_CASE_: Union[str, Any] =['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class __A ( UpperCamelCase__ ): a__ : int = VOCAB_FILES_NAMES a__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : List[str] = PRETRAINED_VOCAB_FILES_MAP a__ : Optional[int] = ["""input_ids""", """attention_mask"""] a__ : Any = MBartTokenizer a__ : List[int] = [] a__ : List[int] = [] def __init__(self : int , __a : Any=None , __a : Dict=None , __a : str="<s>" , __a : Tuple="</s>" , __a : Optional[Any]="</s>" , __a : Dict="<s>" , __a : Optional[int]="<unk>" , __a : int="<pad>" , __a : Any="<mask>" , __a : Dict=None , __a : Tuple=None , __a : Any=None , **__a : List[str] , ): # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ = AddedToken(__a , lstrip=__a , rstrip=__a ) if isinstance(__a , __a ) else mask_token super().__init__( vocab_file=__a , tokenizer_file=__a , bos_token=__a , eos_token=__a , sep_token=__a , cls_token=__a , unk_token=__a , pad_token=__a , mask_token=__a , src_lang=__a , tgt_lang=__a , additional_special_tokens=__a , **__a , ) UpperCAmelCase_ = vocab_file UpperCAmelCase_ = False if not self.vocab_file else True UpperCAmelCase_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) UpperCAmelCase_ = { lang_code: self.convert_tokens_to_ids(__a ) for lang_code in FAIRSEQ_LANGUAGE_CODES } UpperCAmelCase_ = src_lang if src_lang is not None else "en_XX" UpperCAmelCase_ = self.convert_tokens_to_ids(self._src_lang ) UpperCAmelCase_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def _lowercase (self : str ): return self._src_lang @src_lang.setter def _lowercase (self : int , __a : str ): UpperCAmelCase_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowercase (self : List[str] , __a : List[int] , __a : Optional[List[int]] = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowercase (self : Any , __a : List[int] , __a : Optional[List[int]] = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowercase (self : Any , __a : int , __a : str , __a : Optional[str] , __a : Optional[str] , **__a : List[str] ): if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCAmelCase_ = src_lang UpperCAmelCase_ = self(__a , add_special_tokens=__a , return_tensors=__a , **__a ) UpperCAmelCase_ = self.convert_tokens_to_ids(__a ) UpperCAmelCase_ = tgt_lang_id return inputs def _lowercase (self : Union[str, Any] , __a : List[str] , __a : str = "en_XX" , __a : Optional[List[str]] = None , __a : str = "ro_RO" , **__a : int , ): UpperCAmelCase_ = src_lang UpperCAmelCase_ = tgt_lang return super().prepare_seqaseq_batch(__a , __a , **__a ) def _lowercase (self : Optional[int] ): return self.set_src_lang_special_tokens(self.src_lang ) def _lowercase (self : Optional[Any] ): return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowercase (self : int , __a : Optional[int] ): UpperCAmelCase_ = self.convert_tokens_to_ids(__a ) UpperCAmelCase_ = [] UpperCAmelCase_ = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def _lowercase (self : Dict , __a : str ): UpperCAmelCase_ = self.convert_tokens_to_ids(__a ) UpperCAmelCase_ = [] UpperCAmelCase_ = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def _lowercase (self : List[str] , __a : str , __a : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(__a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return UpperCAmelCase_ = os.path.join( __a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__a ): copyfile(self.vocab_file , __a ) return (out_vocab_file,)
1
'''simple docstring''' import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE_: Optional[int] =logging.getLogger() SCREAMING_SNAKE_CASE_: Union[str, Any] =logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __A ( UpperCamelCase__ ): def _lowercase (self : Optional[Any] , __a : str ): os.makedirs(__a , exist_ok=__a ) UpperCAmelCase_ = {"source": "What is love ?", "target": "life"} UpperCAmelCase_ = {"train": 12, "val": 2, "test": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: UpperCAmelCase_ = "\n".join([contents[field]] * n_lines[split] ) with open(os.path.join(__a , f"""{split}.{field}""" ) , "w" ) as f: f.write(__a ) def _lowercase (self : Optional[int] , __a : int , __a : str = "pytorch" ): UpperCAmelCase_ = self.get_auto_remove_tmp_dir() UpperCAmelCase_ = os.path.join(__a , "output" ) UpperCAmelCase_ = os.path.join(__a , "data" ) self._create_dummy_data(data_dir=__a ) UpperCAmelCase_ = f""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(f"""--gpus={gpus}""" ) if is_apex_available(): testargs.append("--fp16" ) else: testargs.append("--gpus=0" ) testargs.append("--distributed_backend=ddp_cpu" ) testargs.append("--num_processes=2" ) UpperCAmelCase_ = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(__a , env=self.get_env() ) UpperCAmelCase_ = os.path.join(__a , "metrics.json" ) with open(__a ) as f: UpperCAmelCase_ = json.load(__a ) return result @require_torch_gpu def _lowercase (self : Optional[int] ): UpperCAmelCase_ = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu def _lowercase (self : Dict ): UpperCAmelCase_ = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_gpu @require_ray def _lowercase (self : Optional[int] ): UpperCAmelCase_ = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu @require_ray def _lowercase (self : Any ): UpperCAmelCase_ = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 )
1
1
import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) __snake_case :Optional[int] = [ '''cross_validation.py''', '''gradient_accumulation.py''', '''local_sgd.py''', '''multi_process_metrics.py''', '''memory.py''', '''automatic_gradient_accumulation.py''', '''fsdp_with_peak_mem_tracking.py''', '''deepspeed_with_config_support.py''', '''megatron_lm_gpt_pretraining.py''', ] class _A ( unittest.TestCase ): def _lowerCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : bool , __SCREAMING_SNAKE_CASE : str = None , __SCREAMING_SNAKE_CASE : list = None): '''simple docstring''' __a = None __a = os.path.abspath(os.path.join('''examples''' , '''by_feature''')) __a = os.path.abspath('''examples''') for item in os.listdir(__SCREAMING_SNAKE_CASE): if item not in EXCLUDE_EXAMPLES: __a = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) if os.path.isfile(__SCREAMING_SNAKE_CASE) and ".py" in item_path: with self.subTest( tested_script=__SCREAMING_SNAKE_CASE , feature_script=__SCREAMING_SNAKE_CASE , tested_section='''main()''' if parser_only else '''training_function()''' , ): __a = compare_against_test( os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) __a = '''\n'''.join(__SCREAMING_SNAKE_CASE) if special_strings is not None: for string in special_strings: __a = diff.replace(__SCREAMING_SNAKE_CASE , '''''') self.assertEqual(__SCREAMING_SNAKE_CASE , '''''') def _lowerCamelCase ( self : Tuple): '''simple docstring''' self.one_complete_example('''complete_nlp_example.py''' , __SCREAMING_SNAKE_CASE) self.one_complete_example('''complete_nlp_example.py''' , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = os.path.abspath(os.path.join('''examples''' , '''cv_example.py''')) __a = [ ''' ''' * 16 + '''{\n\n''', ''' ''' * 20 + '''"accuracy": eval_metric["accuracy"],\n\n''', ''' ''' * 20 + '''"f1": eval_metric["f1"],\n\n''', ''' ''' * 20 + '''"train_loss": total_loss.item() / len(train_dataloader),\n\n''', ''' ''' * 20 + '''"epoch": epoch,\n\n''', ''' ''' * 16 + '''},\n\n''', ''' ''' * 16 + '''step=epoch,\n''', ''' ''' * 12, ''' ''' * 8 + '''for step, batch in enumerate(active_dataloader):\n''', ] self.one_complete_example('''complete_cv_example.py''' , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) self.one_complete_example('''complete_cv_example.py''' , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) @mock.patch.dict(os.environ ,{'''TESTING_MOCKED_DATALOADERS''': '''1'''} ) class _A ( __UpperCAmelCase ): UpperCamelCase__ : List[str] = False @classmethod def _lowerCamelCase ( cls : int): '''simple docstring''' super().setUpClass() __a = tempfile.mkdtemp() __a = os.path.join(cls._tmpdir , '''default_config.yml''') write_basic_config(save_location=cls.configPath) __a = ['''accelerate''', '''launch''', '''--config_file''', cls.configPath] @classmethod def _lowerCamelCase ( cls : Dict): '''simple docstring''' super().tearDownClass() shutil.rmtree(cls._tmpdir) def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = F'\n examples/by_feature/checkpointing.py\n --checkpointing_steps epoch\n --output_dir {self.tmpdir}\n '.split() run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , '''epoch_0'''))) def _lowerCamelCase ( self : str): '''simple docstring''' __a = F'\n examples/by_feature/checkpointing.py\n --checkpointing_steps 1\n --output_dir {self.tmpdir}\n '.split() __a = run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , '''step_2'''))) def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = F'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "epoch_0")}\n '.split() __a = run_command(self._launch_args + testargs , return_stdout=__SCREAMING_SNAKE_CASE) self.assertNotIn('''epoch 0:''' , __SCREAMING_SNAKE_CASE) self.assertIn('''epoch 1:''' , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = F'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "step_2")}\n '.split() __a = run_command(self._launch_args + testargs , return_stdout=__SCREAMING_SNAKE_CASE) if torch.cuda.is_available(): __a = torch.cuda.device_count() else: __a = 1 if num_processes > 1: self.assertNotIn('''epoch 0:''' , __SCREAMING_SNAKE_CASE) self.assertIn('''epoch 1:''' , __SCREAMING_SNAKE_CASE) else: self.assertIn('''epoch 0:''' , __SCREAMING_SNAKE_CASE) self.assertIn('''epoch 1:''' , __SCREAMING_SNAKE_CASE) @slow def _lowerCamelCase ( self : int): '''simple docstring''' __a = ''' examples/by_feature/cross_validation.py --num_folds 2 '''.split() with mock.patch.dict(os.environ , {'''TESTING_MOCKED_DATALOADERS''': '''0'''}): __a = run_command(self._launch_args + testargs , return_stdout=__SCREAMING_SNAKE_CASE) __a = re.findall('''({.+})''' , __SCREAMING_SNAKE_CASE) __a = [r for r in results if '''accuracy''' in r][-1] __a = ast.literal_eval(__SCREAMING_SNAKE_CASE) self.assertGreaterEqual(results['''accuracy'''] , 0.75) def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = ['''examples/by_feature/multi_process_metrics.py'''] run_command(self._launch_args + testargs) @require_trackers @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''}) def _lowerCamelCase ( self : int): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: __a = F'\n examples/by_feature/tracking.py\n --with_tracking\n --project_dir {tmpdir}\n '.split() run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(__SCREAMING_SNAKE_CASE , '''tracking'''))) def _lowerCamelCase ( self : List[Any]): '''simple docstring''' __a = ['''examples/by_feature/gradient_accumulation.py'''] run_command(self._launch_args + testargs) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __a = ['''examples/by_feature/local_sgd.py'''] run_command(self._launch_args + testargs)
131
def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): return base * power(_UpperCAmelCase , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('''Raise base to the power of exponent using recursion...''') __snake_case :List[Any] = int(input('''Enter the base: ''').strip()) __snake_case :Dict = int(input('''Enter the exponent: ''').strip()) __snake_case :int = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents __snake_case :Optional[Any] = 1 / result print(f'{base} to the power of {exponent} is {result}')
131
1
'''simple docstring''' import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def UpperCamelCase_ ( snake_case_ : Union[str, Any] , snake_case_ : List[str] , snake_case_ : Optional[Any]=0 ) -> Tuple: '''simple docstring''' if name is None: __lowerCAmelCase = None else: __lowerCAmelCase = '.' * max(0 , spaces - 2 ) + '# {:' + str(50 - spaces ) + 's}' __lowerCAmelCase = fmt.format(_UpperCAmelCase ) # Print and recurse (if needed). if isinstance(_UpperCAmelCase , _UpperCAmelCase ): if msg is not None: print(_UpperCAmelCase ) for k in val.keys(): recursive_print(_UpperCAmelCase , val[k] , spaces + 2 ) elif isinstance(_UpperCAmelCase , torch.Tensor ): print(_UpperCAmelCase , """:""" , val.size() ) else: print(_UpperCAmelCase , """:""" , _UpperCAmelCase ) def UpperCamelCase_ ( snake_case_ : int , snake_case_ : Union[str, Any] , snake_case_ : Dict , snake_case_ : Optional[Any] , snake_case_ : Optional[int] ) -> str: '''simple docstring''' __lowerCAmelCase = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] __lowerCAmelCase = (num_heads, hidden_size, num_splits) + input_shape[1:] __lowerCAmelCase = param.view(*_UpperCAmelCase ) __lowerCAmelCase = param.transpose(0 , 2 ) __lowerCAmelCase = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] __lowerCAmelCase = (num_heads, num_splits, hidden_size) + input_shape[1:] __lowerCAmelCase = param.view(*_UpperCAmelCase ) __lowerCAmelCase = param.transpose(0 , 1 ).contiguous() __lowerCAmelCase = param.view(*_UpperCAmelCase ) return param def UpperCamelCase_ ( snake_case_ : int , snake_case_ : Any , snake_case_ : Tuple ) -> Tuple: '''simple docstring''' __lowerCAmelCase = {} # old versions did not store training args __lowerCAmelCase = input_state_dict.get("""args""" , _UpperCAmelCase ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) __lowerCAmelCase = ds_args.padded_vocab_size __lowerCAmelCase = ds_args.max_position_embeddings __lowerCAmelCase = ds_args.hidden_size __lowerCAmelCase = ds_args.num_layers __lowerCAmelCase = ds_args.num_attention_heads __lowerCAmelCase = ds_args.ffn_hidden_size # pprint(config) # The number of heads. __lowerCAmelCase = config.n_head # The hidden_size per head. __lowerCAmelCase = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): __lowerCAmelCase = input_state_dict['checkpoint_version'] else: __lowerCAmelCase = 0.0 # The model. __lowerCAmelCase = input_state_dict['model'] # The language model. __lowerCAmelCase = model['language_model'] # The embeddings. __lowerCAmelCase = lm['embedding'] # The word embeddings. __lowerCAmelCase = embeddings['word_embeddings']['weight'] # Truncate the embedding table to vocab_size rows. __lowerCAmelCase = word_embeddings[: config.vocab_size, :] __lowerCAmelCase = word_embeddings # The position embeddings. __lowerCAmelCase = embeddings['position_embeddings']['weight'] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] __lowerCAmelCase = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( f"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match""" ) # Store the position embeddings. __lowerCAmelCase = pos_embeddings # The transformer. __lowerCAmelCase = lm['transformer'] if 'transformer' in lm.keys() else lm['encoder'] # The regex to extract layer names. __lowerCAmelCase = re.compile(r"""layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)""" ) # The simple map of names for "automated" rules. __lowerCAmelCase = { 'attention.dense': '.attn.c_proj.', 'self_attention.dense': '.attn.c_proj.', 'mlp.dense_h_to_4h': '.mlp.c_fc.', 'mlp.dense_4h_to_h': '.mlp.c_proj.', } # Extract the layers. for key, val in transformer.items(): # Match the name. __lowerCAmelCase = layer_re.match(_UpperCAmelCase ) # Stop if that's not a layer if m is None: break # The index of the layer. __lowerCAmelCase = int(m.group(1 ) ) # The name of the operation. __lowerCAmelCase = m.group(2 ) # Is it a weight or a bias? __lowerCAmelCase = m.group(3 ) # The name of the layer. __lowerCAmelCase = f"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith("""layernorm""" ): __lowerCAmelCase = 'ln_1' if op_name.startswith("""input""" ) else 'ln_2' __lowerCAmelCase = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. __lowerCAmelCase = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , _UpperCAmelCase , _UpperCAmelCase ) __lowerCAmelCase = causal_mask # Insert a "dummy" tensor for masked_bias. __lowerCAmelCase = torch.tensor(-1E4 , dtype=torch.floataa ) __lowerCAmelCase = masked_bias __lowerCAmelCase = fix_query_key_value_ordering(_UpperCAmelCase , _UpperCAmelCase , 3 , _UpperCAmelCase , _UpperCAmelCase ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. __lowerCAmelCase = out_val.transpose(0 , 1 ).contiguous() # Store. __lowerCAmelCase = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": __lowerCAmelCase = fix_query_key_value_ordering(_UpperCAmelCase , _UpperCAmelCase , 3 , _UpperCAmelCase , _UpperCAmelCase ) # Store. No change of shape. __lowerCAmelCase = out_val # Transpose the weights. elif weight_or_bias == "weight": __lowerCAmelCase = megatron_to_transformers[op_name] __lowerCAmelCase = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": __lowerCAmelCase = megatron_to_transformers[op_name] __lowerCAmelCase = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. __lowerCAmelCase = transformer['final_layernorm.weight'] __lowerCAmelCase = transformer['final_layernorm.bias'] # For LM head, transformers' wants the matrix to weight embeddings. __lowerCAmelCase = word_embeddings # It should be done! return output_state_dict def UpperCamelCase_ ( ) -> int: '''simple docstring''' __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument("""--print-checkpoint-structure""" , action="""store_true""" ) parser.add_argument( """path_to_checkpoint""" , type=_UpperCAmelCase , help="""Path to the checkpoint file (.zip archive or direct .pt file)""" , ) parser.add_argument( """--config_file""" , default="""""" , type=_UpperCAmelCase , help="""An optional config json file describing the pre-trained model.""" , ) __lowerCAmelCase = parser.parse_args() # Extract the basename. __lowerCAmelCase = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith(""".zip""" ): with zipfile.ZipFile(args.path_to_checkpoint , """r""" ) as checkpoint: with checkpoint.open("""release/mp_rank_00/model_optim_rng.pt""" ) as pytorch_dict: __lowerCAmelCase = torch.load(_UpperCAmelCase , map_location="""cpu""" ) else: __lowerCAmelCase = torch.load(args.path_to_checkpoint , map_location="""cpu""" ) __lowerCAmelCase = input_state_dict.get("""args""" , _UpperCAmelCase ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: __lowerCAmelCase = 'gelu_fast' elif ds_args.openai_gelu: __lowerCAmelCase = 'gelu_new' else: __lowerCAmelCase = 'gelu' else: # in the very early days this used to be "gelu_new" __lowerCAmelCase = 'gelu_new' # Spell out all parameters in case the defaults change. __lowerCAmelCase = GPTaConfig( vocab_size=5_02_57 , n_positions=10_24 , n_embd=10_24 , n_layer=24 , n_head=16 , n_inner=40_96 , activation_function=_UpperCAmelCase , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1E-5 , initializer_range=0.0_2 , summary_type="""cls_index""" , summary_use_proj=_UpperCAmelCase , summary_activation=_UpperCAmelCase , summary_proj_to_labels=_UpperCAmelCase , summary_first_dropout=0.1 , scale_attn_weights=_UpperCAmelCase , use_cache=_UpperCAmelCase , bos_token_id=5_02_56 , eos_token_id=5_02_56 , ) else: __lowerCAmelCase = GPTaConfig.from_json_file(args.config_file ) __lowerCAmelCase = ['GPT2LMHeadModel'] # Convert. print("""Converting""" ) __lowerCAmelCase = convert_megatron_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(_UpperCAmelCase , _UpperCAmelCase ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: __lowerCAmelCase = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": __lowerCAmelCase = 'gpt2' elif tokenizer_type == "PretrainedFromHF": __lowerCAmelCase = ds_args.tokenizer_name_or_path else: raise ValueError(f"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: __lowerCAmelCase = 'gpt2' __lowerCAmelCase = AutoTokenizer.from_pretrained(_UpperCAmelCase ) __lowerCAmelCase = type(_UpperCAmelCase ).__name__ __lowerCAmelCase = tokenizer_class # Store the config to file. print("""Saving config""" ) config.save_pretrained(_UpperCAmelCase ) # Save tokenizer based on args print(f"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(_UpperCAmelCase ) # Store the state_dict to file. __lowerCAmelCase = os.path.join(_UpperCAmelCase , """pytorch_model.bin""" ) print(f"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(_UpperCAmelCase , _UpperCAmelCase ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
229
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase: str = logging.get_logger(__name__) _UpperCamelCase: Optional[Any] = torch.device('cpu') def lowercase__ ( ) -> str: '''simple docstring''' lowercase : Any = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowercase : Any = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im def lowercase__ ( _UpperCAmelCase ) -> int: '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703e00, 2.1107e00, -2.0811e00, 8.8685e-01, 2.4360e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636e-01, 2.3478e-01, -1.6963e00, -1.7381e00, -8.6337e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768e-01, -4.7429e-01, -1.0897e00, -1.0248e00, 3.5523e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330e-01, 2.4211e-01, -6.0185e-01, -8.2789e-01, -6.0446e-02] ) def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: '''simple docstring''' lowercase : Optional[Any] = dct.pop(_UpperCAmelCase ) lowercase : Dict = val def lowercase__ ( _UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowercase : int = [] for k in state_dict.keys(): lowercase : List[Any] = k if ".pwconv" in k: lowercase : Union[str, Any] = k_new.replace('.pwconv' , '.point_wise_conv' ) if ".dwconv" in k: lowercase : Optional[int] = k_new.replace('.dwconv' , '.depth_wise_conv' ) if ".Proj." in k: lowercase : List[Any] = k_new.replace('.Proj.' , '.proj.' ) if "patch_embed" in k_new: lowercase : str = k_new.replace('patch_embed' , 'swiftformer.patch_embed.patch_embedding' ) if "network" in k_new: lowercase : str = k_new.split('.' ) if ls[2].isdigit(): lowercase : Optional[Any] = 'swiftformer.encoder.network.' + ls[1] + '.blocks.' + ls[2] + '.' + '.'.join(ls[3:] ) else: lowercase : Dict = k_new.replace('network' , 'swiftformer.encoder.network' ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: '''simple docstring''' lowercase : Any = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size lowercase : Optional[int] = 10_00 lowercase : Optional[int] = 'huggingface/label-files' lowercase : Optional[int] = 'imagenet-1k-id2label.json' lowercase : str = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) lowercase : Tuple = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowercase : Optional[int] = idalabel lowercase : str = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": lowercase : List[str] = [3, 3, 6, 4] lowercase : int = [48, 56, 1_12, 2_20] elif swiftformer_name == "swiftformer_s": lowercase : str = [3, 3, 9, 6] lowercase : Union[str, Any] = [48, 64, 1_68, 2_24] elif swiftformer_name == "swiftformer_l1": lowercase : Dict = [4, 3, 10, 5] lowercase : Tuple = [48, 96, 1_92, 3_84] elif swiftformer_name == "swiftformer_l3": lowercase : Union[str, Any] = [4, 4, 12, 6] lowercase : str = [64, 1_28, 3_20, 5_12] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith('https' ): lowercase : Optional[int] = torch.hub.load_state_dict_from_url(_UpperCAmelCase , map_location='cpu' , check_hash=_UpperCAmelCase ) else: lowercase : Union[str, Any] = torch.load(_UpperCAmelCase , map_location='cpu' ) lowercase : List[Any] = checkpoint lowercase : Tuple = create_rename_keys(_UpperCAmelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # load HuggingFace model lowercase : Optional[Any] = SwiftFormerForImageClassification(_UpperCAmelCase ).eval() hf_model.load_state_dict(_UpperCAmelCase ) # prepare test inputs lowercase : str = prepare_img() lowercase : Tuple = ViTImageProcessor.from_pretrained('preprocessor_config' ) lowercase : Tuple = processor(images=_UpperCAmelCase , return_tensors='pt' ) # compare outputs from both models lowercase : Dict = get_expected_output(_UpperCAmelCase ) lowercase : Union[str, Any] = hf_model(inputs['pixel_values'] ).logits assert hf_logits.shape == torch.Size([1, 10_00] ) assert torch.allclose(hf_logits[0, 0:5] , _UpperCAmelCase , atol=1e-3 ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f'''Saving model {swiftformer_name} to {pytorch_dump_folder_path}''' ) hf_model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": _UpperCamelCase: Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _UpperCamelCase: Union[str, Any] = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
53
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _UpperCamelCase: Any = logging.get_logger(__name__) class a__ ( SCREAMING_SNAKE_CASE__ ): _lowerCamelCase = ['pixel_values'] def __init__( self : Tuple, lowerCAmelCase : bool = True, lowerCAmelCase : Union[int, float] = 1 / 255, lowerCAmelCase : bool = True, lowerCAmelCase : int = 8, **lowerCAmelCase : Optional[int], ) -> None: super().__init__(**lowerCAmelCase ) lowercase : Dict = do_rescale lowercase : Tuple = rescale_factor lowercase : List[str] = do_pad lowercase : int = pad_size def lowercase ( self : List[Any], lowerCAmelCase : np.ndarray, lowerCAmelCase : float, lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None, **lowerCAmelCase : int ) -> np.ndarray: return rescale(lowerCAmelCase, scale=lowerCAmelCase, data_format=lowerCAmelCase, **lowerCAmelCase ) def lowercase ( self : Union[str, Any], lowerCAmelCase : np.ndarray, lowerCAmelCase : int, lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None ) -> List[Any]: lowercase , lowercase : Tuple = get_image_size(lowerCAmelCase ) lowercase : Optional[Any] = (old_height // size + 1) * size - old_height lowercase : Dict = (old_width // size + 1) * size - old_width return pad(lowerCAmelCase, ((0, pad_height), (0, pad_width)), mode='symmetric', data_format=lowerCAmelCase ) def lowercase ( self : Any, lowerCAmelCase : ImageInput, lowerCAmelCase : Optional[bool] = None, lowerCAmelCase : Optional[float] = None, lowerCAmelCase : Optional[bool] = None, lowerCAmelCase : Optional[int] = None, lowerCAmelCase : Optional[Union[str, TensorType]] = None, lowerCAmelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST, **lowerCAmelCase : Any, ) -> List[Any]: lowercase : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale lowercase : Any = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase : Any = do_pad if do_pad is not None else self.do_pad lowercase : int = pad_size if pad_size is not None else self.pad_size lowercase : Tuple = make_list_of_images(lowerCAmelCase ) if not valid_images(lowerCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) # All transformations expect numpy arrays. lowercase : Dict = [to_numpy_array(lowerCAmelCase ) for image in images] if do_rescale: lowercase : Optional[int] = [self.rescale(image=lowerCAmelCase, scale=lowerCAmelCase ) for image in images] if do_pad: lowercase : List[str] = [self.pad(lowerCAmelCase, size=lowerCAmelCase ) for image in images] lowercase : Optional[int] = [to_channel_dimension_format(lowerCAmelCase, lowerCAmelCase ) for image in images] lowercase : Tuple = {'pixel_values': images} return BatchFeature(data=lowerCAmelCase, tensor_type=lowerCAmelCase )
53
1
'''simple docstring''' import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : List[str] , __a : List[Any] , __a : str=13 , __a : Dict=10 , __a : Tuple=3 , __a : Dict=2 , __a : Any=2 , __a : Union[str, Any]=True , __a : str=True , __a : Optional[int]=32 , __a : List[str]=5 , __a : int=4 , __a : Any=37 , __a : Optional[Any]="gelu" , __a : Optional[int]=0.1 , __a : List[str]=0.1 , __a : Tuple=10 , __a : Union[str, Any]=0.02 , __a : Union[str, Any]="divided_space_time" , __a : Dict=None , ): _a = parent _a = batch_size _a = image_size _a = num_channels _a = patch_size _a = num_frames _a = is_training _a = use_labels _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = intermediate_size _a = hidden_act _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = attention_type _a = initializer_range _a = scope _a = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token _a = (image_size // patch_size) ** 2 _a = (num_frames) * self.num_patches_per_frame + 1 def UpperCamelCase__ ( self : Union[str, Any] ): _a = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.num_labels ) _a = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self : int ): _a = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) _a = self.num_labels return config def UpperCamelCase__ ( self : List[Any] , __a : str , __a : Union[str, Any] , __a : List[str] ): _a = TimesformerModel(config=__a ) model.to(__a ) model.eval() _a = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self : Any , __a : Union[str, Any] , __a : str , __a : List[Any] ): _a = TimesformerForVideoClassification(__a ) model.to(__a ) model.eval() _a = model(__a ) # verify the logits shape _a = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , __a ) def UpperCamelCase__ ( self : Any ): _a = self.prepare_config_and_inputs() _a , _a , _a = config_and_inputs _a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE (lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __a =(TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () __a =( {'feature-extraction': TimesformerModel, 'video-classification': TimesformerForVideoClassification} if is_torch_available() else {} ) __a =False __a =False __a =False __a =False def UpperCamelCase__ ( self : Dict ): _a = TimesformerModelTester(self ) _a = ConfigTester( self , config_class=__a , has_text_modality=__a , hidden_size=37 ) def UpperCamelCase__ ( self : int , __a : List[Any] , __a : List[Any] , __a : Any=False ): _a = copy.deepcopy(__a ) if return_labels: if model_class in get_values(__a ): _a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a ) return inputs_dict def UpperCamelCase__ ( self : int ): self.config_tester.run_common_tests() @unittest.skip(reason="TimeSformer does not use inputs_embeds" ) def UpperCamelCase__ ( self : List[Any] ): pass def UpperCamelCase__ ( self : Optional[int] ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(__a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear ) ) def UpperCamelCase__ ( self : Dict ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(__a ) _a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a = [*signature.parameters.keys()] _a = ["pixel_values"] self.assertListEqual(arg_names[:1] , __a ) def UpperCamelCase__ ( self : Optional[int] ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def UpperCamelCase__ ( self : List[str] ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*__a ) @slow def UpperCamelCase__ ( self : Tuple ): for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = TimesformerModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def UpperCamelCase__ ( self : Optional[Any] ): if not self.has_attentions: pass else: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = True for model_class in self.all_model_classes: _a = self.model_tester.seq_length _a = self.model_tester.num_frames _a = True _a = False _a = True _a = model_class(__a ) model.to(__a ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(__a , __a ) ) _a = outputs.attentions self.assertEqual(len(__a ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _a = True _a = model_class(__a ) model.to(__a ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(__a , __a ) ) _a = outputs.attentions self.assertEqual(len(__a ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) _a = len(__a ) # Check attention is always last and order is fine _a = True _a = True _a = model_class(__a ) model.to(__a ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(__a , __a ) ) self.assertEqual(out_len + 1 , len(__a ) ) _a = outputs.attentions self.assertEqual(len(__a ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def UpperCamelCase__ ( self : str ): def check_hidden_states_output(__a : Union[str, Any] , __a : Optional[Any] , __a : Any ): _a = model_class(__a ) model.to(__a ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(__a , __a ) ) _a = outputs.hidden_states _a = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(__a ) , __a ) _a = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = True check_hidden_states_output(__a , __a , __a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a = True check_hidden_states_output(__a , __a , __a ) def _lowerCamelCase ( ) -> int: _a = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename="eating_spaghetti.npy" , repo_type="dataset" ) _a = np.load(lowercase ) return list(lowercase ) @require_torch @require_vision class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" @cached_property def UpperCamelCase__ ( self : List[str] ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def UpperCamelCase__ ( self : str ): _a = TimesformerForVideoClassification.from_pretrained("facebook/timesformer-base-finetuned-k400" ).to( __a ) _a = self.default_image_processor _a = prepare_video() _a = image_processor(video[:8] , return_tensors="pt" ).to(__a ) # forward pass with torch.no_grad(): _a = model(**__a ) # verify the logits _a = torch.Size((1, 4_00) ) self.assertEqual(outputs.logits.shape , __a ) _a = torch.tensor([-0.3016, -0.7713, -0.4205] ).to(__a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4 ) )
63
from __future__ import annotations from functools import lru_cache from math import ceil __lowerCamelCase : str = 100 __lowerCamelCase : Any = set(range(3, NUM_PRIMES, 2)) primes.add(2) __lowerCamelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def A_ ( _lowerCAmelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} UpperCamelCase : set[int] = set() UpperCamelCase : int UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def A_ ( _lowerCAmelCase = 5000 ) -> int | None: for number_to_partition in range(1 , _lowerCAmelCase ): if len(partition(_lowerCAmelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
52
0
'''simple docstring''' from __future__ import annotations def _lowerCAmelCase ( __snake_case : int , __snake_case : int ) -> tuple[int, int]: if b == 0: return (1, 0) ((__A) ,(__A)) : Tuple = extended_euclid(__snake_case , a % b ) __A : Dict = a // b return (y, x - k * y) def _lowerCAmelCase ( __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int ) -> int: ((__A) ,(__A)) : Optional[Any] = extended_euclid(__snake_case , __snake_case ) __A : List[str] = na * na __A : List[str] = ra * x * na + ra * y * na return (n % m + m) % m def _lowerCAmelCase ( __snake_case : int , __snake_case : int ) -> int: ((__A) ,(__A)) : Tuple = extended_euclid(__snake_case , __snake_case ) if b < 0: __A : Optional[int] = (b % n + n) % n return b def _lowerCAmelCase ( __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int ) -> int: __A ,__A : int = invert_modulo(__snake_case , __snake_case ), invert_modulo(__snake_case , __snake_case ) __A : List[Any] = na * na __A : str = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name='''chinese_remainder_theorem''', verbose=True) testmod(name='''chinese_remainder_theorem2''', verbose=True) testmod(name='''invert_modulo''', verbose=True) testmod(name='''extended_euclid''', verbose=True)
190
'''simple docstring''' import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class SCREAMING_SNAKE_CASE (a__ ): def __get__( self , _UpperCAmelCase , _UpperCAmelCase=None): '''simple docstring''' if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute') __A : Optional[Any] = '__cached_' + self.fget.__name__ __A : Union[str, Any] = getattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase) if cached is None: __A : int = self.fget(_UpperCAmelCase) setattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase) return cached def _lowerCAmelCase ( __snake_case : Optional[Any] ) -> Tuple: __A : Dict = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'invalid truth value {val!r}' ) def _lowerCAmelCase ( __snake_case : int ) -> Tuple: if is_torch_fx_proxy(__snake_case ): return True if is_torch_available(): import torch if isinstance(__snake_case , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(__snake_case , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(__snake_case , (jnp.ndarray, Tracer) ): return True return isinstance(__snake_case , np.ndarray ) def _lowerCAmelCase ( __snake_case : Dict ) -> List[str]: return isinstance(__snake_case , np.ndarray ) def _lowerCAmelCase ( __snake_case : Optional[int] ) -> Any: return _is_numpy(__snake_case ) def _lowerCAmelCase ( __snake_case : int ) -> Union[str, Any]: import torch return isinstance(__snake_case , torch.Tensor ) def _lowerCAmelCase ( __snake_case : str ) -> Optional[int]: return False if not is_torch_available() else _is_torch(__snake_case ) def _lowerCAmelCase ( __snake_case : Any ) -> List[Any]: import torch return isinstance(__snake_case , torch.device ) def _lowerCAmelCase ( __snake_case : List[Any] ) -> Dict: return False if not is_torch_available() else _is_torch_device(__snake_case ) def _lowerCAmelCase ( __snake_case : Union[str, Any] ) -> Optional[int]: import torch if isinstance(__snake_case , __snake_case ): if hasattr(__snake_case , __snake_case ): __A : str = getattr(__snake_case , __snake_case ) else: return False return isinstance(__snake_case , torch.dtype ) def _lowerCAmelCase ( __snake_case : Optional[Any] ) -> Any: return False if not is_torch_available() else _is_torch_dtype(__snake_case ) def _lowerCAmelCase ( __snake_case : List[str] ) -> int: import tensorflow as tf return isinstance(__snake_case , tf.Tensor ) def _lowerCAmelCase ( __snake_case : Dict ) -> List[Any]: return False if not is_tf_available() else _is_tensorflow(__snake_case ) def _lowerCAmelCase ( __snake_case : str ) -> List[str]: import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(__snake_case , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(__snake_case ) return type(__snake_case ) == tf.Tensor def _lowerCAmelCase ( __snake_case : Dict ) -> Tuple: return False if not is_tf_available() else _is_tf_symbolic_tensor(__snake_case ) def _lowerCAmelCase ( __snake_case : int ) -> Union[str, Any]: import jax.numpy as jnp # noqa: F811 return isinstance(__snake_case , jnp.ndarray ) def _lowerCAmelCase ( __snake_case : int ) -> List[str]: return False if not is_flax_available() else _is_jax(__snake_case ) def _lowerCAmelCase ( __snake_case : Optional[Any] ) -> Dict: if isinstance(__snake_case , (dict, UserDict) ): return {k: to_py_obj(__snake_case ) for k, v in obj.items()} elif isinstance(__snake_case , (list, tuple) ): return [to_py_obj(__snake_case ) for o in obj] elif is_tf_tensor(__snake_case ): return obj.numpy().tolist() elif is_torch_tensor(__snake_case ): return obj.detach().cpu().tolist() elif is_jax_tensor(__snake_case ): return np.asarray(__snake_case ).tolist() elif isinstance(__snake_case , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def _lowerCAmelCase ( __snake_case : Tuple ) -> Optional[int]: if isinstance(__snake_case , (dict, UserDict) ): return {k: to_numpy(__snake_case ) for k, v in obj.items()} elif isinstance(__snake_case , (list, tuple) ): return np.array(__snake_case ) elif is_tf_tensor(__snake_case ): return obj.numpy() elif is_torch_tensor(__snake_case ): return obj.detach().cpu().numpy() elif is_jax_tensor(__snake_case ): return np.asarray(__snake_case ) else: return obj class SCREAMING_SNAKE_CASE (a__ ): def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = fields(self) # Safety and consistency checks if not len(_UpperCAmelCase): raise ValueError(F'{self.__class__.__name__} has no fields.') if not all(field.default is None for field in class_fields[1:]): raise ValueError(F'{self.__class__.__name__} should not have more than one required field.') __A : Tuple = getattr(self , class_fields[0].name) __A : Tuple = all(getattr(self , field.name) is None for field in class_fields[1:]) if other_fields_are_none and not is_tensor(_UpperCAmelCase): if isinstance(_UpperCAmelCase , _UpperCAmelCase): __A : List[str] = first_field.items() __A : List[Any] = True else: try: __A : List[Any] = iter(_UpperCAmelCase) __A : Optional[Any] = True except TypeError: __A : List[Any] = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(_UpperCAmelCase): if ( not isinstance(_UpperCAmelCase , (list, tuple)) or not len(_UpperCAmelCase) == 2 or not isinstance(element[0] , _UpperCAmelCase) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute __A : Optional[int] = first_field else: # If we have a mixed iterator, raise an error raise ValueError( F'Cannot set key/value for {element}. It needs to be a tuple (key, value).') break setattr(self , element[0] , element[1]) if element[1] is not None: __A : Optional[int] = element[1] elif first_field is not None: __A : Dict = first_field else: for field in class_fields: __A : List[str] = getattr(self , field.name) if v is not None: __A : Union[str, Any] = v def __delitem__( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' raise Exception(F'You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.') def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' raise Exception(F'You cannot use ``setdefault`` on a {self.__class__.__name__} instance.') def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' raise Exception(F'You cannot use ``pop`` on a {self.__class__.__name__} instance.') def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' raise Exception(F'You cannot use ``update`` on a {self.__class__.__name__} instance.') def __getitem__( self , _UpperCAmelCase): '''simple docstring''' if isinstance(_UpperCAmelCase , _UpperCAmelCase): __A : List[Any] = dict(self.items()) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(_UpperCAmelCase , _UpperCAmelCase) super().__setattr__(_UpperCAmelCase , _UpperCAmelCase) def __setitem__( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' super().__setitem__(_UpperCAmelCase , _UpperCAmelCase) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(_UpperCAmelCase , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return tuple(self[k] for k in self.keys()) class SCREAMING_SNAKE_CASE (a__ , a__ ): @classmethod def SCREAMING_SNAKE_CASE ( cls , _UpperCAmelCase): '''simple docstring''' raise ValueError( F'{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys())}') class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''longest''' lowerCAmelCase = '''max_length''' lowerCAmelCase = '''do_not_pad''' class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''pt''' lowerCAmelCase = '''tf''' lowerCAmelCase = '''np''' lowerCAmelCase = '''jax''' class SCREAMING_SNAKE_CASE : def __init__( self , _UpperCAmelCase): '''simple docstring''' __A : Optional[Any] = context_managers __A : Optional[int] = ExitStack() def __enter__( self): '''simple docstring''' for context_manager in self.context_managers: self.stack.enter_context(_UpperCAmelCase) def __exit__( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' self.stack.__exit__(*_UpperCAmelCase , **_UpperCAmelCase) def _lowerCAmelCase ( __snake_case : List[str] ) -> int: __A : Any = infer_framework(__snake_case ) if framework == "tf": __A : int = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": __A : Any = inspect.signature(model_class.forward ) # PyTorch models else: __A : Union[str, Any] = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def _lowerCAmelCase ( __snake_case : int ) -> List[Any]: __A : Any = model_class.__name__ __A : Optional[int] = infer_framework(__snake_case ) if framework == "tf": __A : List[str] = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": __A : List[Any] = inspect.signature(model_class.forward ) # PyTorch models else: __A : List[Any] = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def _lowerCAmelCase ( __snake_case : MutableMapping , __snake_case : str = "" , __snake_case : str = "." ) -> Union[str, Any]: def _flatten_dict(__snake_case : Tuple , __snake_case : List[Any]="" , __snake_case : Tuple="." ): for k, v in d.items(): __A : List[Any] = str(__snake_case ) + delimiter + str(__snake_case ) if parent_key else k if v and isinstance(__snake_case , __snake_case ): yield from flatten_dict(__snake_case , __snake_case , delimiter=__snake_case ).items() else: yield key, v return dict(_flatten_dict(__snake_case , __snake_case , __snake_case ) ) @contextmanager def _lowerCAmelCase ( __snake_case : Any , __snake_case : bool = False ) -> List[str]: if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def _lowerCAmelCase ( __snake_case : Optional[Any] , __snake_case : Optional[Any]=None ) -> int: if is_numpy_array(__snake_case ): return np.transpose(__snake_case , axes=__snake_case ) elif is_torch_tensor(__snake_case ): return array.T if axes is None else array.permute(*__snake_case ) elif is_tf_tensor(__snake_case ): import tensorflow as tf return tf.transpose(__snake_case , perm=__snake_case ) elif is_jax_tensor(__snake_case ): return jnp.transpose(__snake_case , axes=__snake_case ) else: raise ValueError(f'Type not supported for transpose: {type(__snake_case )}.' ) def _lowerCAmelCase ( __snake_case : List[Any] , __snake_case : str ) -> str: if is_numpy_array(__snake_case ): return np.reshape(__snake_case , __snake_case ) elif is_torch_tensor(__snake_case ): return array.reshape(*__snake_case ) elif is_tf_tensor(__snake_case ): import tensorflow as tf return tf.reshape(__snake_case , __snake_case ) elif is_jax_tensor(__snake_case ): return jnp.reshape(__snake_case , __snake_case ) else: raise ValueError(f'Type not supported for reshape: {type(__snake_case )}.' ) def _lowerCAmelCase ( __snake_case : Tuple , __snake_case : List[str]=None ) -> Any: if is_numpy_array(__snake_case ): return np.squeeze(__snake_case , axis=__snake_case ) elif is_torch_tensor(__snake_case ): return array.squeeze() if axis is None else array.squeeze(dim=__snake_case ) elif is_tf_tensor(__snake_case ): import tensorflow as tf return tf.squeeze(__snake_case , axis=__snake_case ) elif is_jax_tensor(__snake_case ): return jnp.squeeze(__snake_case , axis=__snake_case ) else: raise ValueError(f'Type not supported for squeeze: {type(__snake_case )}.' ) def _lowerCAmelCase ( __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: if is_numpy_array(__snake_case ): return np.expand_dims(__snake_case , __snake_case ) elif is_torch_tensor(__snake_case ): return array.unsqueeze(dim=__snake_case ) elif is_tf_tensor(__snake_case ): import tensorflow as tf return tf.expand_dims(__snake_case , axis=__snake_case ) elif is_jax_tensor(__snake_case ): return jnp.expand_dims(__snake_case , axis=__snake_case ) else: raise ValueError(f'Type not supported for expand_dims: {type(__snake_case )}.' ) def _lowerCAmelCase ( __snake_case : Optional[Any] ) -> Dict: if is_numpy_array(__snake_case ): return np.size(__snake_case ) elif is_torch_tensor(__snake_case ): return array.numel() elif is_tf_tensor(__snake_case ): import tensorflow as tf return tf.size(__snake_case ) elif is_jax_tensor(__snake_case ): return array.size else: raise ValueError(f'Type not supported for expand_dims: {type(__snake_case )}.' ) def _lowerCAmelCase ( __snake_case : str , __snake_case : Tuple ) -> Union[str, Any]: for key, value in auto_map.items(): if isinstance(__snake_case , (tuple, list) ): __A : Tuple = [f'{repo_id}--{v}' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: __A : Dict = f'{repo_id}--{value}' return auto_map def _lowerCAmelCase ( __snake_case : List[str] ) -> int: for base_class in inspect.getmro(__snake_case ): __A : int = base_class.__module__ __A : List[str] = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'Could not infer framework from class {model_class}.' )
190
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase__ = logging.get_logger(__name__) def snake_case_ ( A_ : Dict ): '''simple docstring''' if isinstance(A_, (list, tuple) ) and isinstance(videos[0], (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(A_, (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(A_ ): return [[videos]] raise ValueError(F'''Could not make batched video from {videos}''' ) class __snake_case ( _lowercase): snake_case__ : Dict = ["pixel_values"] def __init__( self : str , __lowerCAmelCase : bool = True , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : PILImageResampling = PILImageResampling.BILINEAR , __lowerCAmelCase : bool = True , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : bool = True , __lowerCAmelCase : Union[int, float] = 1 / 2_5_5 , __lowerCAmelCase : bool = True , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , **__lowerCAmelCase : Tuple , ): """simple docstring""" super().__init__(**__lowerCAmelCase ) _lowerCamelCase : Optional[int] = size if size is not None else {'''shortest_edge''': 2_2_4} _lowerCamelCase : List[Any] = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} _lowerCamelCase : str = get_size_dict(__lowerCAmelCase , param_name='''crop_size''' ) _lowerCamelCase : int = do_resize _lowerCamelCase : Optional[Any] = size _lowerCamelCase : Optional[int] = do_center_crop _lowerCamelCase : Optional[Any] = crop_size _lowerCamelCase : str = resample _lowerCamelCase : Any = do_rescale _lowerCamelCase : List[Any] = rescale_factor _lowerCamelCase : List[Any] = do_normalize _lowerCamelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCamelCase : str = image_std if image_std is not None else IMAGENET_STANDARD_STD def SCREAMING_SNAKE_CASE ( self : Dict , __lowerCAmelCase : np.ndarray , __lowerCAmelCase : Dict[str, int] , __lowerCAmelCase : PILImageResampling = PILImageResampling.BILINEAR , __lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCAmelCase : List[Any] , ): """simple docstring""" _lowerCamelCase : Optional[Any] = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) if "shortest_edge" in size: _lowerCamelCase : Tuple = get_resize_output_image_size(__lowerCAmelCase , size['''shortest_edge'''] , default_to_square=__lowerCAmelCase ) elif "height" in size and "width" in size: _lowerCamelCase : str = (size['''height'''], size['''width''']) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(__lowerCAmelCase , size=__lowerCAmelCase , resample=__lowerCAmelCase , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : np.ndarray , __lowerCAmelCase : Dict[str, int] , __lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCAmelCase : Dict , ): """simple docstring""" _lowerCamelCase : int = get_size_dict(__lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(__lowerCAmelCase , size=(size['''height'''], size['''width''']) , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : np.ndarray , __lowerCAmelCase : Union[int, float] , __lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCAmelCase : Dict , ): """simple docstring""" return rescale(__lowerCAmelCase , scale=__lowerCAmelCase , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : np.ndarray , __lowerCAmelCase : Union[float, List[float]] , __lowerCAmelCase : Union[float, List[float]] , __lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCAmelCase : Tuple , ): """simple docstring""" return normalize(__lowerCAmelCase , mean=__lowerCAmelCase , std=__lowerCAmelCase , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : ImageInput , __lowerCAmelCase : bool = None , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : PILImageResampling = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : float = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , __lowerCAmelCase : Optional[ChannelDimension] = ChannelDimension.FIRST , ): """simple docstring""" if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. _lowerCamelCase : Any = to_numpy_array(__lowerCAmelCase ) if do_resize: _lowerCamelCase : Tuple = self.resize(image=__lowerCAmelCase , size=__lowerCAmelCase , resample=__lowerCAmelCase ) if do_center_crop: _lowerCamelCase : List[str] = self.center_crop(__lowerCAmelCase , size=__lowerCAmelCase ) if do_rescale: _lowerCamelCase : Optional[Any] = self.rescale(image=__lowerCAmelCase , scale=__lowerCAmelCase ) if do_normalize: _lowerCamelCase : int = self.normalize(image=__lowerCAmelCase , mean=__lowerCAmelCase , std=__lowerCAmelCase ) _lowerCamelCase : str = to_channel_dimension_format(__lowerCAmelCase , __lowerCAmelCase ) return image def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : ImageInput , __lowerCAmelCase : bool = None , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : PILImageResampling = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : Dict[str, int] = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : float = None , __lowerCAmelCase : bool = None , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , __lowerCAmelCase : Optional[Union[float, List[float]]] = None , __lowerCAmelCase : Optional[Union[str, TensorType]] = None , __lowerCAmelCase : ChannelDimension = ChannelDimension.FIRST , **__lowerCAmelCase : Any , ): """simple docstring""" _lowerCamelCase : List[Any] = do_resize if do_resize is not None else self.do_resize _lowerCamelCase : str = resample if resample is not None else self.resample _lowerCamelCase : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCamelCase : Any = do_rescale if do_rescale is not None else self.do_rescale _lowerCamelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCamelCase : List[Any] = do_normalize if do_normalize is not None else self.do_normalize _lowerCamelCase : Dict = image_mean if image_mean is not None else self.image_mean _lowerCamelCase : int = image_std if image_std is not None else self.image_std _lowerCamelCase : List[str] = size if size is not None else self.size _lowerCamelCase : str = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) _lowerCamelCase : List[str] = crop_size if crop_size is not None else self.crop_size _lowerCamelCase : str = get_size_dict(__lowerCAmelCase , param_name='''crop_size''' ) if not valid_images(__lowerCAmelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) _lowerCamelCase : Any = make_batched(__lowerCAmelCase ) _lowerCamelCase : Tuple = [ [ self._preprocess_image( image=__lowerCAmelCase , do_resize=__lowerCAmelCase , size=__lowerCAmelCase , resample=__lowerCAmelCase , do_center_crop=__lowerCAmelCase , crop_size=__lowerCAmelCase , do_rescale=__lowerCAmelCase , rescale_factor=__lowerCAmelCase , do_normalize=__lowerCAmelCase , image_mean=__lowerCAmelCase , image_std=__lowerCAmelCase , data_format=__lowerCAmelCase , ) for img in video ] for video in videos ] _lowerCamelCase : List[str] = {'''pixel_values''': videos} return BatchFeature(data=__lowerCAmelCase , tensor_type=__lowerCAmelCase )
72
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__ = { '''configuration_lxmert''': ['''LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LxmertConfig'''], '''tokenization_lxmert''': ['''LxmertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''LxmertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''LxmertEncoder''', '''LxmertForPreTraining''', '''LxmertForQuestionAnswering''', '''LxmertModel''', '''LxmertPreTrainedModel''', '''LxmertVisualFeatureEncoder''', '''LxmertXLayer''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLxmertForPreTraining''', '''TFLxmertMainLayer''', '''TFLxmertModel''', '''TFLxmertPreTrainedModel''', '''TFLxmertVisualFeatureEncoder''', ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
1
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def UpperCAmelCase_ ( self : Optional[int] ) -> List[Any]: """simple docstring""" snake_case_ : Optional[int] = 'laion/clap-htsat-unfused' snake_case_ : List[Any] = tempfile.mkdtemp() def UpperCAmelCase_ ( self : Union[str, Any] , **_A : Union[str, Any] ) -> Optional[Any]: """simple docstring""" return RobertaTokenizer.from_pretrained(self.checkpoint , **__lowerCamelCase ) def UpperCAmelCase_ ( self : List[Any] , **_A : int ) -> Optional[int]: """simple docstring""" return ClapFeatureExtractor.from_pretrained(self.checkpoint , **__lowerCamelCase ) def UpperCAmelCase_ ( self : List[Any] ) -> int: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" snake_case_ : List[Any] = self.get_tokenizer() snake_case_ : Tuple = self.get_feature_extractor() snake_case_ : List[Any] = ClapProcessor(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) snake_case_ : Optional[Any] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __lowerCamelCase ) def UpperCAmelCase_ ( self : List[str] ) -> Optional[int]: """simple docstring""" snake_case_ : Optional[int] = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) snake_case_ : Union[str, Any] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) snake_case_ : Any = self.get_feature_extractor(do_normalize=__lowerCamelCase , padding_value=1.0 ) snake_case_ : int = ClapProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __lowerCamelCase ) def UpperCAmelCase_ ( self : List[Any] ) -> str: """simple docstring""" snake_case_ : Dict = self.get_feature_extractor() snake_case_ : Tuple = self.get_tokenizer() snake_case_ : str = ClapProcessor(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase ) snake_case_ : str = floats_list((3, 1000) ) snake_case_ : Union[str, Any] = feature_extractor(__lowerCamelCase , return_tensors='np' ) snake_case_ : Optional[Any] = processor(audios=__lowerCamelCase , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase_ ( self : str ) -> str: """simple docstring""" snake_case_ : Dict = self.get_feature_extractor() snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = ClapProcessor(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase ) snake_case_ : Optional[int] = 'This is a test string' snake_case_ : List[str] = processor(text=__lowerCamelCase ) snake_case_ : Dict = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCAmelCase_ ( self : int ) -> Optional[int]: """simple docstring""" snake_case_ : Tuple = self.get_feature_extractor() snake_case_ : Optional[int] = self.get_tokenizer() snake_case_ : Dict = ClapProcessor(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase ) snake_case_ : Tuple = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] snake_case_ : Union[str, Any] = processor.batch_decode(__lowerCamelCase ) snake_case_ : Optional[Any] = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" snake_case_ : List[str] = self.get_feature_extractor() snake_case_ : Tuple = self.get_tokenizer() snake_case_ : Any = ClapProcessor(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg='`processor` and `feature_extractor` model input names do not match' , )
367
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class SCREAMING_SNAKE_CASE_ : __magic_name__: int = MBartConfig __magic_name__: str = {} __magic_name__: Union[str, Any] = "gelu" def __init__( self : List[str] , _A : Optional[int] , _A : List[Any]=13 , _A : List[Any]=7 , _A : Dict=True , _A : Tuple=False , _A : Optional[Any]=99 , _A : Dict=32 , _A : str=2 , _A : str=4 , _A : Tuple=37 , _A : Tuple=0.1 , _A : Union[str, Any]=0.1 , _A : Optional[int]=20 , _A : Dict=2 , _A : List[str]=1 , _A : Union[str, Any]=0 , ) -> List[Any]: """simple docstring""" snake_case_ : str = parent snake_case_ : List[str] = batch_size snake_case_ : List[str] = seq_length snake_case_ : Union[str, Any] = is_training snake_case_ : Optional[int] = use_labels snake_case_ : Dict = vocab_size snake_case_ : Union[str, Any] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Optional[Any] = num_attention_heads snake_case_ : Union[str, Any] = intermediate_size snake_case_ : Any = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : Optional[Any] = max_position_embeddings snake_case_ : Optional[Any] = eos_token_id snake_case_ : Tuple = pad_token_id snake_case_ : int = bos_token_id def UpperCAmelCase_ ( self : List[str] ) -> Tuple: """simple docstring""" snake_case_ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case_ : int = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ : Dict = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : Optional[Any] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case_ : Union[str, Any] = prepare_mbart_inputs_dict(_A , _A , _A ) return config, inputs_dict def UpperCAmelCase_ ( self : Optional[Any] , _A : Optional[Any] , _A : int ) -> str: """simple docstring""" snake_case_ : Dict = TFMBartModel(config=_A ).get_decoder() snake_case_ : Any = inputs_dict['input_ids'] snake_case_ : List[Any] = input_ids[:1, :] snake_case_ : Dict = inputs_dict['attention_mask'][:1, :] snake_case_ : Tuple = inputs_dict['head_mask'] snake_case_ : List[Any] = 1 # first forward pass snake_case_ : Any = model(_A , attention_mask=_A , head_mask=_A , use_cache=_A ) snake_case_ ,snake_case_ : str = outputs.to_tuple() snake_case_ : int = past_key_values[1] def SCREAMING_SNAKE_CASE__ ( __a , __a , __a , __a=None , __a=None , __a=None , __a=None , __a=None , ): if attention_mask is None: snake_case_ : Optional[int] = tf.cast(tf.math.not_equal(__a , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case_ : str = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case_ : str = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ : int = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case_ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class SCREAMING_SNAKE_CASE_ ( snake_case_ , snake_case_ , unittest.TestCase ): __magic_name__: Tuple = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () __magic_name__: int = (TFMBartForConditionalGeneration,) if is_tf_available() else () __magic_name__: Union[str, Any] = ( { "conversational": TFMBartForConditionalGeneration, "feature-extraction": TFMBartModel, "summarization": TFMBartForConditionalGeneration, "text2text-generation": TFMBartForConditionalGeneration, "translation": TFMBartForConditionalGeneration, } if is_tf_available() else {} ) __magic_name__: Tuple = True __magic_name__: Tuple = False __magic_name__: Any = False def UpperCAmelCase_ ( self : Any , _A : Union[str, Any] , _A : List[Any] , _A : str , _A : int , _A : Dict ) -> Union[str, Any]: """simple docstring""" if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCAmelCase_ ( self : Dict ) -> Tuple: """simple docstring""" snake_case_ : Optional[Any] = TFMBartModelTester(self ) snake_case_ : List[Any] = ConfigTester(self , config_class=_A ) def UpperCAmelCase_ ( self : Optional[Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_A ) @require_sentencepiece @require_tokenizers @require_tf class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): __magic_name__: Optional[int] = [ " UN Chief Says There Is No Military Solution in Syria", ] __magic_name__: Union[str, Any] = [ "Şeful ONU declară că nu există o soluţie militară în Siria", ] __magic_name__: List[Any] = "facebook/mbart-large-en-ro" @cached_property def UpperCAmelCase_ ( self : str ) -> List[Any]: """simple docstring""" return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCAmelCase_ ( self : List[Any] ) -> Any: """simple docstring""" snake_case_ : List[str] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def UpperCAmelCase_ ( self : Optional[int] , **_A : str ) -> int: """simple docstring""" snake_case_ : List[str] = self.translate_src_text(**_A ) self.assertListEqual(self.expected_text , _A ) def UpperCAmelCase_ ( self : Union[str, Any] , **_A : Dict ) -> int: """simple docstring""" snake_case_ : Optional[Any] = self.tokenizer(self.src_text , **_A , return_tensors='tf' ) snake_case_ : int = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) snake_case_ : Any = self.tokenizer.batch_decode(_A , skip_special_tokens=_A ) return generated_words @slow def UpperCAmelCase_ ( self : str ) -> List[str]: """simple docstring""" self._assert_generated_batch_equal_expected()
88
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class A_ : def __init__( self , _A , _A=2 , _A=True , _A=False , _A=1_0 , _A=3 , _A=3_2 * 4 , _A=3_2 * 6 , _A=4 , _A=3_2 , ): '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = is_training UpperCAmelCase = use_auxiliary_loss UpperCAmelCase = num_queries UpperCAmelCase = num_channels UpperCAmelCase = min_size UpperCAmelCase = max_size UpperCAmelCase = num_labels UpperCAmelCase = mask_feature_size def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( __snake_case ) UpperCAmelCase = torch.ones([self.batch_size, self.min_size, self.max_size] , device=__snake_case ) UpperCAmelCase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=__snake_case ) > 0.5 ).float() UpperCAmelCase = (torch.rand((self.batch_size, self.num_labels) , device=__snake_case ) > 0.5).long() UpperCAmelCase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _lowercase ( self ): '''simple docstring''' return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_2_8 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def _lowercase ( self , _A , _A ): '''simple docstring''' UpperCAmelCase = output.encoder_hidden_states UpperCAmelCase = output.pixel_decoder_hidden_states UpperCAmelCase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(__snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__snake_case ) , config.decoder_config.decoder_layers ) def _lowercase ( self , _A , _A , _A , _A=False ): '''simple docstring''' with torch.no_grad(): UpperCAmelCase = MaskFormerModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase = model(pixel_values=__snake_case , pixel_mask=__snake_case ) UpperCAmelCase = model(__snake_case , output_hidden_states=__snake_case ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(__snake_case , __snake_case ) def _lowercase ( self , _A , _A , _A , _A , _A ): '''simple docstring''' UpperCAmelCase = MaskFormerForInstanceSegmentation(config=__snake_case ) model.to(__snake_case ) model.eval() def comm_check_on_output(_A ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase = model(pixel_values=__snake_case , pixel_mask=__snake_case ) UpperCAmelCase = model(__snake_case ) comm_check_on_output(__snake_case ) UpperCAmelCase = model( pixel_values=__snake_case , pixel_mask=__snake_case , mask_labels=__snake_case , class_labels=__snake_case ) comm_check_on_output(__snake_case ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class A_ (snake_case_ , snake_case_ , unittest.TestCase ): UpperCAmelCase__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () UpperCAmelCase__ = ( {'''feature-extraction''': MaskFormerModel, '''image-segmentation''': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = MaskFormerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__snake_case , has_text_modality=__snake_case ) def _lowercase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__snake_case , **__snake_case , output_hidden_states=__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*__snake_case ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def _lowercase ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _lowercase ( self ): '''simple docstring''' pass def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__snake_case ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __snake_case ) @slow def _lowercase ( self ): '''simple docstring''' for model_name in ["facebook/maskformer-swin-small-coco"]: UpperCAmelCase = MaskFormerModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = (self.model_tester.min_size,) * 2 UpperCAmelCase = { '''pixel_values''': torch.randn((2, 3, *size) , device=__snake_case ), '''mask_labels''': torch.randn((2, 1_0, *size) , device=__snake_case ), '''class_labels''': torch.zeros(2 , 1_0 , device=__snake_case ).long(), } UpperCAmelCase = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(__snake_case ) UpperCAmelCase = model(**__snake_case ) self.assertTrue(outputs.loss is not None ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__snake_case , **__snake_case , output_hidden_states=__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__snake_case ).to(__snake_case ) UpperCAmelCase = model(**__snake_case , output_attentions=__snake_case ) self.assertTrue(outputs.attentions is not None ) def _lowercase ( self ): '''simple docstring''' if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss UpperCAmelCase = self.all_model_classes[1] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() UpperCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() UpperCAmelCase = model(__snake_case , mask_labels=__snake_case , class_labels=__snake_case ).loss loss.backward() def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.all_model_classes[1] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() UpperCAmelCase = model(__snake_case , mask_labels=__snake_case , class_labels=__snake_case ) UpperCAmelCase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't UpperCAmelCase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=__snake_case ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __A : int = 1E-4 def __SCREAMING_SNAKE_CASE ( ) -> int: '''simple docstring''' UpperCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class A_ (unittest.TestCase ): @cached_property def _lowercase ( self ): '''simple docstring''' return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(__snake_case ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) UpperCAmelCase = torch.tensor( [[-0.04_82, 0.92_28, 0.49_51], [-0.25_47, 0.80_17, 0.85_27], [-0.00_69, 0.33_85, -0.00_89]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) UpperCAmelCase = torch.tensor( [[-0.84_22, -0.84_34, -0.97_18], [-1.01_44, -0.55_65, -0.41_95], [-1.00_38, -0.44_84, -0.19_61]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) UpperCAmelCase = torch.tensor( [[0.28_52, -0.01_59, 0.97_35], [0.62_54, 0.18_58, 0.85_29], [-0.06_80, -0.41_16, 1.84_13]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) # masks_queries_logits UpperCAmelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) UpperCAmelCase = [ [-1.3_73_71_24, -1.7_72_49_37, -1.9_36_42_33], [-1.5_97_72_81, -1.9_86_79_39, -2.1_52_36_95], [-1.5_79_53_98, -1.9_26_98_32, -2.09_39_42], ] UpperCAmelCase = torch.tensor(__snake_case ).to(__snake_case ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) # class_queries_logits UpperCAmelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase = torch.tensor( [ [1.6512E00, -5.2572E00, -3.3519E00], [3.6169E-02, -5.9025E00, -2.9313E00], [1.0766E-04, -7.7630E00, -5.1263E00], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) # masks_queries_logits UpperCAmelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) UpperCAmelCase = [[-0.90_46, -2.63_66, -4.60_62], [-3.41_79, -5.78_90, -8.80_57], [-4.91_79, -7.65_60, -10.77_11]] UpperCAmelCase = torch.tensor(__snake_case ).to(__snake_case ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) # class_queries_logits UpperCAmelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase = torch.tensor( [[4.71_88, -3.25_85, -2.88_57], [6.68_71, -2.91_81, -1.24_87], [7.24_49, -2.27_64, -2.18_74]] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = image_processor( [np.zeros((3, 8_0_0, 1_3_3_3) ), np.zeros((3, 8_0_0, 1_3_3_3) )] , segmentation_maps=[np.zeros((3_8_4, 3_8_4) ).astype(np.floataa ), np.zeros((3_8_4, 3_8_4) ).astype(np.floataa )] , return_tensors='''pt''' , ) UpperCAmelCase = inputs['''pixel_values'''].to(__snake_case ) UpperCAmelCase = [el.to(__snake_case ) for el in inputs['''mask_labels''']] UpperCAmelCase = [el.to(__snake_case ) for el in inputs['''class_labels''']] with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) self.assertTrue(outputs.loss is not None )
273
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class UpperCAmelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = StableDiffusionControlNetImgaImgPipeline _lowercase: Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) _lowercase: Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self : List[str] ) -> List[str]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Any , __snake_case : str , __snake_case : Any=0 ) -> str: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ) _lowerCAmelCase = floats_tensor(control_image.shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : Optional[int] ) -> List[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = StableDiffusionControlNetImgaImgPipeline _lowercase: Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Any = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__snake_case : Optional[Any] ): if isinstance(__snake_case , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = MultiControlNetModel([controlneta, controlneta] ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Tuple , __snake_case : int , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), ] _lowerCAmelCase = floats_tensor(control_image[0].shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : List[str] ) -> Dict: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) _lowerCAmelCase = 10.0 _lowerCAmelCase = 4 _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def lowercase__ ( self : int ) -> str: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Optional[Any] ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : int ) -> str: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__snake_case ) except NotImplementedError: pass @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Union[str, Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) _lowerCAmelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__snake_case , controlnet=__snake_case ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase = """evil space-punk bird""" _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = pipe( __snake_case , __snake_case , control_image=__snake_case , generator=__snake_case , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) _lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
70
0
from timeit import timeit def __lowercase ( __lowerCAmelCase : int ): """simple docstring""" if number < 0: raise ValueError('the value of input must not be negative' ) a__ = 0 while number: number &= number - 1 result += 1 return result def __lowercase ( __lowerCAmelCase : int ): """simple docstring""" if number < 0: raise ValueError('the value of input must not be negative' ) a__ = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def __lowercase ( ): """simple docstring""" def do_benchmark(__lowerCAmelCase : int ) -> None: a__ = 'import __main__ as z' print(F'Benchmark when {number = }:' ) print(F'{get_set_bits_count_using_modulo_operator(__lowerCAmelCase ) = }' ) a__ = timeit('z.get_set_bits_count_using_modulo_operator(25)' , setup=__lowerCAmelCase ) print(F'timeit() runs in {timing} seconds' ) print(F'{get_set_bits_count_using_brian_kernighans_algorithm(__lowerCAmelCase ) = }' ) a__ = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)' , setup=__lowerCAmelCase , ) print(F'timeit() runs in {timing} seconds' ) for number in (2_5, 3_7, 5_8, 0): do_benchmark(__lowerCAmelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
353
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : Optional[int] = logging.get_logger(__name__) snake_case : Union[str, Any] = { '''uclanlp/visualbert-vqa''': '''https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json''', '''uclanlp/visualbert-vqa-pre''': '''https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json''', '''uclanlp/visualbert-vqa-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-vcr''': '''https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json''', '''uclanlp/visualbert-vcr-pre''': '''https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json''', '''uclanlp/visualbert-vcr-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-nlvr2''': '''https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-pre''': '''https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json''' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class snake_case_ (lowerCamelCase_ ): UpperCAmelCase__ : Dict = '''visual_bert''' def __init__( self :Optional[int] ,__snake_case :Any=3_05_22 ,__snake_case :str=7_68 ,__snake_case :Any=5_12 ,__snake_case :Any=12 ,__snake_case :int=12 ,__snake_case :str=30_72 ,__snake_case :int="gelu" ,__snake_case :Optional[int]=0.1 ,__snake_case :str=0.1 ,__snake_case :Union[str, Any]=5_12 ,__snake_case :Tuple=2 ,__snake_case :Union[str, Any]=0.02 ,__snake_case :Optional[Any]=1E-12 ,__snake_case :Optional[Any]=False ,__snake_case :int=True ,__snake_case :Any=1 ,__snake_case :Optional[int]=0 ,__snake_case :Tuple=2 ,**__snake_case :Any ,) -> Union[str, Any]: super().__init__(pad_token_id=__snake_case ,bos_token_id=__snake_case ,eos_token_id=__snake_case ,**__snake_case ) a__ = vocab_size a__ = max_position_embeddings a__ = hidden_size a__ = visual_embedding_dim a__ = num_hidden_layers a__ = num_attention_heads a__ = intermediate_size a__ = hidden_act a__ = hidden_dropout_prob a__ = attention_probs_dropout_prob a__ = initializer_range a__ = type_vocab_size a__ = layer_norm_eps a__ = bypass_transformer a__ = special_visual_initialize
109
0
"""simple docstring""" import argparse import os import re A: List[Any] = "src/transformers" # Pattern that looks at the indentation in a line. A: List[str] = re.compile(R"^(\s*)\S") # Pattern that matches `"key":" and puts `key` in group 0. A: str = re.compile(R"^\s*\"([^\"]+)\":") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. A: Tuple = re.compile(R"^\s*_import_structure\[\"([^\"]+)\"\]") # Pattern that matches `"key",` and puts `key` in group 0. A: Optional[Any] = re.compile(R"^\s*\"([^\"]+)\",\s*$") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. A: str = re.compile(R"\[([^\]]+)\]") def _snake_case ( UpperCamelCase : Optional[int] ): UpperCAmelCase : int = _re_indent.search(UpperCamelCase ) return "" if search is None else search.groups()[0] def _snake_case ( UpperCamelCase : Optional[int] , UpperCamelCase : str="" , UpperCamelCase : List[Any]=None , UpperCamelCase : Optional[Any]=None ): UpperCAmelCase : Any = 0 UpperCAmelCase : int = code.split("""\n""" ) if start_prompt is not None: while not lines[index].startswith(UpperCamelCase ): index += 1 UpperCAmelCase : Tuple = ["""\n""".join(lines[:index] )] else: UpperCAmelCase : Any = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). UpperCAmelCase : Dict = [lines[index]] index += 1 while index < len(UpperCamelCase ) and (end_prompt is None or not lines[index].startswith(UpperCamelCase )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(UpperCamelCase ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + """ """ ): current_block.append(lines[index] ) blocks.append("""\n""".join(UpperCamelCase ) ) if index < len(UpperCamelCase ) - 1: UpperCAmelCase : Optional[int] = [lines[index + 1]] index += 1 else: UpperCAmelCase : Tuple = [] else: blocks.append("""\n""".join(UpperCamelCase ) ) UpperCAmelCase : Tuple = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(UpperCamelCase ) > 0: blocks.append("""\n""".join(UpperCamelCase ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(UpperCamelCase ): blocks.append("""\n""".join(lines[index:] ) ) return blocks def _snake_case ( UpperCamelCase : Optional[Any] ): def _inner(UpperCamelCase : Optional[Any] ): return key(UpperCamelCase ).lower().replace("""_""" , """""" ) return _inner def _snake_case ( UpperCamelCase : str , UpperCamelCase : Optional[int]=None ): # If no key is provided, we use a noop. def noop(UpperCamelCase : List[str] ): return x if key is None: UpperCAmelCase : Dict = noop # Constants are all uppercase, they go first. UpperCAmelCase : List[Any] = [obj for obj in objects if key(UpperCamelCase ).isupper()] # Classes are not all uppercase but start with a capital, they go second. UpperCAmelCase : Dict = [obj for obj in objects if key(UpperCamelCase )[0].isupper() and not key(UpperCamelCase ).isupper()] # Functions begin with a lowercase, they go last. UpperCAmelCase : Any = [obj for obj in objects if not key(UpperCamelCase )[0].isupper()] UpperCAmelCase : str = ignore_underscore(UpperCamelCase ) return sorted(UpperCamelCase , key=UpperCamelCase ) + sorted(UpperCamelCase , key=UpperCamelCase ) + sorted(UpperCamelCase , key=UpperCamelCase ) def _snake_case ( UpperCamelCase : str ): # This inner function sort imports between [ ]. def _replace(UpperCamelCase : int ): UpperCAmelCase : List[Any] = match.groups()[0] if "," not in imports: return F"[{imports}]" UpperCAmelCase : Optional[int] = [part.strip().replace("""\"""" , """""" ) for part in imports.split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase : Union[str, Any] = keys[:-1] return "[" + ", ".join([F"\"{k}\"" for k in sort_objects(UpperCamelCase )] ) + "]" UpperCAmelCase : Optional[Any] = import_statement.split("""\n""" ) if len(UpperCamelCase ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. UpperCAmelCase : Union[str, Any] = 2 if lines[1].strip() == """[""" else 1 UpperCAmelCase : str = [(i, _re_strip_line.search(UpperCamelCase ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] UpperCAmelCase : Union[str, Any] = sort_objects(UpperCamelCase , key=lambda UpperCamelCase : x[1] ) UpperCAmelCase : Any = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(UpperCamelCase ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: UpperCAmelCase : Optional[Any] = _re_bracket_content.sub(_replace , lines[1] ) else: UpperCAmelCase : Union[str, Any] = [part.strip().replace("""\"""" , """""" ) for part in lines[1].split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase : int = keys[:-1] UpperCAmelCase : List[Any] = get_indent(lines[1] ) + """, """.join([F"\"{k}\"" for k in sort_objects(UpperCamelCase )] ) return "\n".join(UpperCamelCase ) else: # Finally we have to deal with imports fitting on one line UpperCAmelCase : Dict = _re_bracket_content.sub(_replace , UpperCamelCase ) return import_statement def _snake_case ( UpperCamelCase : List[Any] , UpperCamelCase : Tuple=True ): with open(UpperCamelCase , encoding="""utf-8""" ) as f: UpperCAmelCase : Optional[int] = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 UpperCAmelCase : Union[str, Any] = split_code_in_indented_blocks( UpperCamelCase , start_prompt="""_import_structure = {""" , end_prompt="""if TYPE_CHECKING:""" ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(UpperCamelCase ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. UpperCAmelCase : List[str] = main_blocks[block_idx] UpperCAmelCase : Dict = block.split("""\n""" ) # Get to the start of the imports. UpperCAmelCase : Tuple = 0 while line_idx < len(UpperCamelCase ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: UpperCAmelCase : Any = len(UpperCamelCase ) else: line_idx += 1 if line_idx >= len(UpperCamelCase ): continue # Ignore beginning and last line: they don't contain anything. UpperCAmelCase : List[str] = """\n""".join(block_lines[line_idx:-1] ) UpperCAmelCase : Optional[int] = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. UpperCAmelCase : List[Any] = split_code_in_indented_blocks(UpperCamelCase , indent_level=UpperCamelCase ) # We have two categories of import key: list or _import_structure[key].append/extend UpperCAmelCase : Optional[Any] = _re_direct_key if """_import_structure = {""" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. UpperCAmelCase : Optional[int] = [(pattern.search(UpperCamelCase ).groups()[0] if pattern.search(UpperCamelCase ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. UpperCAmelCase : Dict = [(i, key) for i, key in enumerate(UpperCamelCase ) if key is not None] UpperCAmelCase : List[Any] = [x[0] for x in sorted(UpperCamelCase , key=lambda UpperCamelCase : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. UpperCAmelCase : Dict = 0 UpperCAmelCase : Dict = [] for i in range(len(UpperCamelCase ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: UpperCAmelCase : Tuple = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(UpperCamelCase ) count += 1 # And we put our main block back together with its first and last line. UpperCAmelCase : Optional[int] = """\n""".join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(UpperCamelCase ): if check_only: return True else: print(F"Overwriting {file}." ) with open(UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write("""\n""".join(UpperCamelCase ) ) def _snake_case ( UpperCamelCase : Union[str, Any]=True ): UpperCAmelCase : str = [] for root, _, files in os.walk(UpperCamelCase ): if "__init__.py" in files: UpperCAmelCase : Optional[Any] = sort_imports(os.path.join(UpperCamelCase , """__init__.py""" ) , check_only=UpperCamelCase ) if result: UpperCAmelCase : Union[str, Any] = [os.path.join(UpperCamelCase , """__init__.py""" )] if len(UpperCamelCase ) > 0: raise ValueError(F"Would overwrite {len(UpperCamelCase )} files, run `make style`." ) if __name__ == "__main__": A: Optional[Any] = argparse.ArgumentParser() parser.add_argument("--check_only", action="store_true", help="Whether to only check or fix style.") A: int = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
109
"""simple docstring""" import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def _snake_case ( UpperCamelCase : Dataset , UpperCamelCase : Dict[str, str] ): UpperCAmelCase : Any = args.log_outputs UpperCAmelCase : Any = """_""".join(args.dataset.split("""/""" ) + [args.config, args.split] ) # load metric UpperCAmelCase : List[Any] = load_metric("""wer""" ) UpperCAmelCase : Any = load_metric("""cer""" ) # compute metrics UpperCAmelCase : int = wer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) UpperCAmelCase : str = cer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) # print & log results UpperCAmelCase : Tuple = F"WER: {wer_result}\nCER: {cer_result}" print(UpperCamelCase ) with open(F"{dataset_id}_eval_results.txt" , """w""" ) as f: f.write(UpperCamelCase ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: UpperCAmelCase : str = F"log_{dataset_id}_predictions.txt" UpperCAmelCase : Tuple = F"log_{dataset_id}_targets.txt" with open(UpperCamelCase , """w""" ) as p, open(UpperCamelCase , """w""" ) as t: # mapping function to write output def write_to_file(UpperCamelCase : List[Any] , UpperCamelCase : List[Any] ): p.write(F"{i}" + """\n""" ) p.write(batch["""prediction"""] + """\n""" ) t.write(F"{i}" + """\n""" ) t.write(batch["""target"""] + """\n""" ) result.map(UpperCamelCase , with_indices=UpperCamelCase ) def _snake_case ( UpperCamelCase : str ): UpperCAmelCase : List[str] = """[,?.!\-\;\:\"“%‘”�—’…–]""" # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training UpperCAmelCase : Dict = re.sub(UpperCamelCase , """""" , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! UpperCAmelCase : List[str] = ["""\n\n""", """\n""", """ """, """ """] for t in token_sequences_to_ignore: UpperCAmelCase : Optional[Any] = """ """.join(text.split(UpperCamelCase ) ) return text def _snake_case ( UpperCamelCase : Tuple ): # load dataset UpperCAmelCase : Union[str, Any] = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=UpperCamelCase ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor UpperCAmelCase : Optional[int] = AutoFeatureExtractor.from_pretrained(args.model_id ) UpperCAmelCase : Any = feature_extractor.sampling_rate # resample audio UpperCAmelCase : List[str] = dataset.cast_column("""audio""" , Audio(sampling_rate=UpperCamelCase ) ) # load eval pipeline if args.device is None: UpperCAmelCase : Optional[int] = 0 if torch.cuda.is_available() else -1 UpperCAmelCase : Tuple = pipeline("""automatic-speech-recognition""" , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(UpperCamelCase : Any ): UpperCAmelCase : Any = asr( batch["""audio"""]["""array"""] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) UpperCAmelCase : Tuple = prediction["""text"""] UpperCAmelCase : List[str] = normalize_text(batch["""sentence"""] ) return batch # run inference on all examples UpperCAmelCase : int = dataset.map(UpperCamelCase , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": A: List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_id", type=str, required=True, help="Model identifier. Should be loadable with 🤗 Transformers" ) parser.add_argument( "--dataset", type=str, required=True, help="Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets", ) parser.add_argument( "--config", type=str, required=True, help="Config of the dataset. *E.g.* `'en'` for Common Voice" ) parser.add_argument("--split", type=str, required=True, help="Split of the dataset. *E.g.* `'test'`") parser.add_argument( "--chunk_length_s", type=float, default=None, help="Chunk length in seconds. Defaults to 5 seconds." ) parser.add_argument( "--stride_length_s", type=float, default=None, help="Stride of the audio chunks. Defaults to 1 second." ) parser.add_argument( "--log_outputs", action="store_true", help="If defined, write outputs to log file for analysis." ) parser.add_argument( "--device", type=int, default=None, help="The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.", ) A: Union[str, Any] = parser.parse_args() main(args)
109
1
'''simple docstring''' import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _lowerCamelCase ( *lowercase : Optional[Any] , lowercase : Tuple = None , lowercase : List[Any]=True , lowercase : Optional[Any]=2 ) -> Optional[Any]: from .. import __version__ _a = take_from _a = () if not isinstance(args[0] , a__ ): _a = (args,) for attribute, version_name, message in args: if version.parse(version.parse(a__ ).base_version ) >= version.parse(a__ ): raise ValueError( F'The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers\'' F' version {__version__} is >= {version_name}' ) _a = None if isinstance(a__ , a__ ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(a__ ),) _a = F'The `{attribute}` argument is deprecated and will be removed in version {version_name}.' elif hasattr(a__ , a__ ): values += (getattr(a__ , a__ ),) _a = F'The `{attribute}` attribute is deprecated and will be removed in version {version_name}.' elif deprecated_kwargs is None: _a = F'`{attribute}` is deprecated and will be removed in version {version_name}.' if warning is not None: _a = warning + ''' ''' if standard_warn else '''''' warnings.warn(warning + message , a__ , stacklevel=a__ ) if isinstance(a__ , a__ ) and len(a__ ) > 0: _a = inspect.getouterframes(inspect.currentframe() )[1] _a = call_frame.filename _a = call_frame.lineno _a = call_frame.function _a = next(iter(deprecated_kwargs.items() ) ) raise TypeError(F'{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`' ) if len(a__ ) == 0: return elif len(a__ ) == 1: return values[0] return values
369
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __SCREAMING_SNAKE_CASE (lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" @register_to_config def __init__( self : List[Any] , __a : int , __a : int , __a : int , __a : float , __a : int , __a : int , __a : int , __a : int , __a : str , __a : bool = False , ): super().__init__() _a = nn.Embedding(__a , __a ) _a = nn.Embedding(__a , __a ) _a = False _a = nn.Dropout(p=__a ) _a = TaConfig( vocab_size=__a , d_model=__a , num_heads=__a , d_kv=__a , d_ff=__a , dropout_rate=__a , feed_forward_proj=__a , is_decoder=__a , is_encoder_decoder=__a , ) _a = nn.ModuleList() for lyr_num in range(__a ): _a = TaBlock(__a ) self.encoders.append(__a ) _a = TaLayerNorm(__a ) _a = nn.Dropout(p=__a ) def UpperCamelCase__ ( self : str , __a : Union[str, Any] , __a : Dict ): _a = self.token_embedder(__a ) _a = encoder_input_tokens.shape[1] _a = torch.arange(__a , device=encoder_input_tokens.device ) x += self.position_encoding(__a ) _a = self.dropout_pre(__a ) # inverted the attention mask _a = encoder_input_tokens.size() _a = self.get_extended_attention_mask(__a , __a ) for lyr in self.encoders: _a = lyr(__a , __a )[0] _a = self.layer_norm(__a ) return self.dropout_post(__a ), encoder_inputs_mask
346
0
import unittest import numpy as np def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , ) -> np.ndarray: lowerCAmelCase = np.shape(snake_case__ ) lowerCAmelCase = np.shape(snake_case__ ) lowerCAmelCase = np.shape(snake_case__ ) if shape_a[0] != shape_b[0]: lowerCAmelCase = ( '''Expected the same number of rows for A and B. ''' f"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(snake_case__ ) if shape_b[1] != shape_c[1]: lowerCAmelCase = ( '''Expected the same number of columns for B and C. ''' f"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(snake_case__ ) lowerCAmelCase = pseudo_inv if a_inv is None: try: lowerCAmelCase = np.linalg.inv(snake_case__ ) except np.linalg.LinAlgError: raise ValueError( '''Input matrix A is not invertible. Cannot compute Schur complement.''' ) return mat_c - mat_b.T @ a_inv @ mat_b class lowercase_ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1], [6, 3]] ) lowerCAmelCase = schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.block([[a, b], [b.T, c]] ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = np.linalg.det(__SCREAMING_SNAKE_CASE ) self.assertAlmostEqual(__SCREAMING_SNAKE_CASE , det_a * det_s ) def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self ) ->None: lowerCAmelCase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCAmelCase = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCAmelCase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): schur_complement(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
338
lowercase__ : Optional[int] = '''ABCDEFGHIJKLMNOPQRSTUVWXYZ''' def SCREAMING_SNAKE_CASE_ ( ) -> None: lowerCAmelCase = input('''Enter message: ''' ) lowerCAmelCase = input('''Enter key [alphanumeric]: ''' ) lowerCAmelCase = input('''Encrypt/Decrypt [e/d]: ''' ) if mode.lower().startswith('''e''' ): lowerCAmelCase = '''encrypt''' lowerCAmelCase = encrypt_message(snake_case__ , snake_case__ ) elif mode.lower().startswith('''d''' ): lowerCAmelCase = '''decrypt''' lowerCAmelCase = decrypt_message(snake_case__ , snake_case__ ) print(f"\n{mode.title()}ed message:" ) print(snake_case__ ) def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ ) -> str: return translate_message(snake_case__ , snake_case__ , '''encrypt''' ) def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ ) -> str: return translate_message(snake_case__ , snake_case__ , '''decrypt''' ) def SCREAMING_SNAKE_CASE_ ( snake_case__ , snake_case__ , snake_case__ ) -> str: lowerCAmelCase = [] lowerCAmelCase = 0 lowerCAmelCase = key.upper() for symbol in message: lowerCAmelCase = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(snake_case__ ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(snake_case__ ): lowerCAmelCase = 0 else: translated.append(snake_case__ ) return "".join(snake_case__ ) if __name__ == "__main__": main()
338
1
import numpy as np A : Optional[Any] = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class lowerCamelCase : """simple docstring""" def __init__( self : List[str] ) -> None: SCREAMING_SNAKE_CASE_ = np.array(__magic_name__ ) def __A ( self : Optional[int] , __magic_name__ : str ) -> np.ndarray: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = np.where(letter == self.SQUARE ) SCREAMING_SNAKE_CASE_ = np.concatenate([indexa + 1, indexa + 1] ) return indexes def __A ( self : int , __magic_name__ : int , __magic_name__ : int ) -> str: SCREAMING_SNAKE_CASE_ = self.SQUARE[indexa - 1, indexa - 1] return letter def __A ( self : Dict , __magic_name__ : str ) -> str: SCREAMING_SNAKE_CASE_ = message.lower() SCREAMING_SNAKE_CASE_ = message.replace(" " , "" ) SCREAMING_SNAKE_CASE_ = message.replace("j" , "i" ) SCREAMING_SNAKE_CASE_ = np.empty((2, len(__magic_name__ )) ) for letter_index in range(len(__magic_name__ ) ): SCREAMING_SNAKE_CASE_ = self.letter_to_numbers(message[letter_index] ) SCREAMING_SNAKE_CASE_ = numbers[0] SCREAMING_SNAKE_CASE_ = numbers[1] SCREAMING_SNAKE_CASE_ = first_step.reshape(2 * len(__magic_name__ ) ) SCREAMING_SNAKE_CASE_ = "" for numbers_index in range(len(__magic_name__ ) ): SCREAMING_SNAKE_CASE_ = int(second_step[numbers_index * 2] ) SCREAMING_SNAKE_CASE_ = int(second_step[(numbers_index * 2) + 1] ) SCREAMING_SNAKE_CASE_ = self.numbers_to_letter(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = encoded_message + letter return encoded_message def __A ( self : Optional[Any] , __magic_name__ : str ) -> str: SCREAMING_SNAKE_CASE_ = message.lower() message.replace(" " , "" ) SCREAMING_SNAKE_CASE_ = np.empty(2 * len(__magic_name__ ) ) for letter_index in range(len(__magic_name__ ) ): SCREAMING_SNAKE_CASE_ = self.letter_to_numbers(message[letter_index] ) SCREAMING_SNAKE_CASE_ = numbers[0] SCREAMING_SNAKE_CASE_ = numbers[1] SCREAMING_SNAKE_CASE_ = first_step.reshape((2, len(__magic_name__ )) ) SCREAMING_SNAKE_CASE_ = "" for numbers_index in range(len(__magic_name__ ) ): SCREAMING_SNAKE_CASE_ = int(second_step[0, numbers_index] ) SCREAMING_SNAKE_CASE_ = int(second_step[1, numbers_index] ) SCREAMING_SNAKE_CASE_ = self.numbers_to_letter(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = decoded_message + letter return decoded_message
305
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input A : Union[str, Any] = "Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine" def a__ ( ): SCREAMING_SNAKE_CASE_ = _ask_options( "In which compute environment are you running?" , ["This machine", "AWS (Amazon SageMaker)"] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: SCREAMING_SNAKE_CASE_ = get_sagemaker_input() else: SCREAMING_SNAKE_CASE_ = get_cluster_input() return config def a__ ( __UpperCamelCase=None ): if subparsers is not None: SCREAMING_SNAKE_CASE_ = subparsers.add_parser("config" , description=__UpperCamelCase ) else: SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser("Accelerate config command" , description=__UpperCamelCase ) parser.add_argument( "--config_file" , default=__UpperCamelCase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=__UpperCamelCase ) return parser def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = get_user_input() if args.config_file is not None: SCREAMING_SNAKE_CASE_ = args.config_file else: if not os.path.isdir(__UpperCamelCase ): os.makedirs(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = default_yaml_config_file if config_file.endswith(".json" ): config.to_json_file(__UpperCamelCase ) else: config.to_yaml_file(__UpperCamelCase ) print(F'''accelerate configuration saved at {config_file}''' ) def a__ ( ): SCREAMING_SNAKE_CASE_ = config_command_parser() SCREAMING_SNAKE_CASE_ = parser.parse_args() config_command(__UpperCamelCase ) if __name__ == "__main__": main()
305
1
from __future__ import annotations from collections import namedtuple def __lowerCAmelCase ( a__ , a__ , a__ ) -> tuple: __a = namedtuple('''result''' , '''name value''' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('''Only one argument must be 0''' ) elif power < 0: raise ValueError( '''Power cannot be negative in any electrical/electronics system''' ) elif voltage == 0: return result('''voltage''' , power / current ) elif current == 0: return result('''current''' , power / voltage ) elif power == 0: return result('''power''' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
6
import torch def UpperCAmelCase__ ( ): if torch.cuda.is_available(): lowercase :Optional[int] = torch.cuda.device_count() else: lowercase :Dict = 0 print(F"Successfully ran on {num_gpus} GPUs" ) if __name__ == "__main__": main()
236
0
"""simple docstring""" import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def UpperCAmelCase ( UpperCAmelCase ) -> Dict: # picklable for multiprocessing return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def UpperCAmelCase ( ) -> List[str]: with parallel_backend('spark' ): assert ParallelBackendConfig.backend_name == "spark" snake_case_ = [1, 2, 3] with pytest.raises(UpperCAmelCase ): with parallel_backend('unsupported backend' ): map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=2 ) with pytest.raises(UpperCAmelCase ): with parallel_backend('unsupported backend' ): map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('num_proc' , [2, -1] ) def UpperCAmelCase ( UpperCAmelCase ) -> Optional[Any]: snake_case_ = [1, 2] snake_case_ = {'a': 1, 'b': 2} snake_case_ = {'a': [1, 2], 'b': [3, 4]} snake_case_ = {'a': {'1': 1}, 'b': 2} snake_case_ = {'a': 1, 'b': 2, 'c': 3, 'd': 4} snake_case_ = [2, 3] snake_case_ = {'a': 2, 'b': 3} snake_case_ = {'a': [2, 3], 'b': [4, 5]} snake_case_ = {'a': {'1': 2}, 'b': 3} snake_case_ = {'a': 2, 'b': 3, 'c': 4, 'd': 5} with parallel_backend('spark' ): assert map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=UpperCAmelCase ) == expected_map_nested_sa assert map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=UpperCAmelCase ) == expected_map_nested_sa assert map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=UpperCAmelCase ) == expected_map_nested_sa assert map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=UpperCAmelCase ) == expected_map_nested_sa assert map_nested(UpperCAmelCase , UpperCAmelCase , num_proc=UpperCAmelCase ) == expected_map_nested_sa
312
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> list[str]: if partitions <= 0: raise ValueError('partitions must be a positive number!' ) if partitions > number_of_bytes: raise ValueError('partitions can not > number_of_bytes!' ) snake_case_ = number_of_bytes // partitions snake_case_ = [] for i in range(UpperCAmelCase ): snake_case_ = i * bytes_per_partition + 1 snake_case_ = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(f'{start_bytes}-{end_bytes}' ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
312
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class _lowercase (a_ ): '''simple docstring''' lowercase__ = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
128
from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) UpperCAmelCase : Tuple =_symbol_database.Default() UpperCAmelCase : List[Any] =_descriptor_pool.Default().AddSerializedFile( b"""\n\x19sentencepiece_model.proto\x12\rsentencepiece\"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12\"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12\"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18\" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse\"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32\".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL\"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03""" ) UpperCAmelCase : Optional[int] =globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, """sentencepiece_model_pb2""", _globals) if _descriptor._USE_C_DESCRIPTORS is False: UpperCAmelCase : str =None UpperCAmelCase : List[Any] =b"""H\003""" # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" UpperCAmelCase : str =45 UpperCAmelCase : Optional[Any] =1581 UpperCAmelCase : Dict =1517 UpperCAmelCase : str =1570 UpperCAmelCase : Optional[int] =1584 UpperCAmelCase : str =1793 UpperCAmelCase : Any =1795 UpperCAmelCase : Dict =1916 UpperCAmelCase : str =1864 UpperCAmelCase : Dict =1905 UpperCAmelCase : Union[str, Any] =1919 UpperCAmelCase : Any =2429 UpperCAmelCase : Dict =2208 UpperCAmelCase : int =2418 UpperCAmelCase : str =2323 UpperCAmelCase : Any =2407 # @@protoc_insertion_point(module_scope)
128
1
'''simple docstring''' from __future__ import annotations from random import random class _A : def __init__( self , __UpperCAmelCase = None ) -> str: '''simple docstring''' __UpperCAmelCase : int = value __UpperCAmelCase : Optional[Any] = random() __UpperCAmelCase : Node | None = None __UpperCAmelCase : Node | None = None def __repr__( self ) -> str: '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} , indent=1 ) def __str__( self ) -> str: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = str(self.value ) + """ """ __UpperCAmelCase : List[str] = str(self.left or """""" ) __UpperCAmelCase : Optional[Any] = str(self.right or """""" ) return value + left + right def lowercase_ ( lowerCAmelCase__ : Node | None , lowerCAmelCase__ : int ): """simple docstring""" if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: __UpperCAmelCase , __UpperCAmelCase : List[str] = split(root.left , lowerCAmelCase__ ) return left, root else: __UpperCAmelCase , __UpperCAmelCase : Tuple = split(root.right , lowerCAmelCase__ ) return root, right def lowercase_ ( lowerCAmelCase__ : Node | None , lowerCAmelCase__ : Node | None ): """simple docstring""" if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: __UpperCAmelCase : Optional[Any] = merge(left.right , lowerCAmelCase__ ) return left else: __UpperCAmelCase : int = merge(lowerCAmelCase__ , right.left ) return right def lowercase_ ( lowerCAmelCase__ : Node | None , lowerCAmelCase__ : int ): """simple docstring""" __UpperCAmelCase : str = Node(lowerCAmelCase__ ) __UpperCAmelCase , __UpperCAmelCase : int = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def lowercase_ ( lowerCAmelCase__ : Node | None , lowerCAmelCase__ : int ): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : List[Any] = split(lowerCAmelCase__ , value - 1 ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def lowercase_ ( lowerCAmelCase__ : Node | None ): """simple docstring""" if not root: # None return else: inorder(root.left ) print(root.value , end=""",""" ) inorder(root.right ) def lowercase_ ( lowerCAmelCase__ : Node | None , lowerCAmelCase__ : str ): """simple docstring""" for arg in args.split(): if arg[0] == "+": __UpperCAmelCase : Dict = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": __UpperCAmelCase : Dict = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print("""Unknown command""" ) return root def lowercase_ ( ): """simple docstring""" __UpperCAmelCase : Any = None print( """enter numbers to create a tree, + value to add value into treap, """ """- value to erase all nodes with value. 'q' to quit. """ ) __UpperCAmelCase : Tuple = input() while args != "q": __UpperCAmelCase : int = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) __UpperCAmelCase : Optional[int] = input() print("""good by!""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
16
'''simple docstring''' import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _A ( unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_MASKED_LM_MAPPING _SCREAMING_SNAKE_CASE : Tuple = TF_MODEL_FOR_MASKED_LM_MAPPING def __A ( self ) -> Any: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __A ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase : List[str] = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , top_k=2 , framework="""tf""" ) __UpperCAmelCase : Union[str, Any] = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ {"""sequence""": """My name is grouped""", """score""": 2.1E-05, """token""": 38_015, """token_str""": """ grouped"""}, {"""sequence""": """My name is accuser""", """score""": 2.1E-05, """token""": 25_506, """token_str""": """ accuser"""}, ] , ) __UpperCAmelCase : List[str] = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ { """sequence""": """The largest city in France is grouped""", """score""": 2.1E-05, """token""": 38_015, """token_str""": """ grouped""", }, { """sequence""": """The largest city in France is accuser""", """score""": 2.1E-05, """token""": 25_506, """token_str""": """ accuser""", }, ] , ) __UpperCAmelCase : Union[str, Any] = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ {"""sequence""": """My name is Clara""", """score""": 2E-05, """token""": 13_606, """token_str""": """ Clara"""}, {"""sequence""": """My name is Patrick""", """score""": 2E-05, """token""": 3_499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Te""", """score""": 1.9E-05, """token""": 2_941, """token_str""": """ Te"""}, ] , ) @require_torch def __A ( self ) -> Dict: '''simple docstring''' __UpperCAmelCase : Dict = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , top_k=2 , framework="""pt""" ) __UpperCAmelCase : Union[str, Any] = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ {"""sequence""": """My name is Maul""", """score""": 2.2E-05, """token""": 35_676, """token_str""": """ Maul"""}, {"""sequence""": """My name isELS""", """score""": 2.2E-05, """token""": 16_416, """token_str""": """ELS"""}, ] , ) __UpperCAmelCase : Dict = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ { """sequence""": """The largest city in France is Maul""", """score""": 2.2E-05, """token""": 35_676, """token_str""": """ Maul""", }, {"""sequence""": """The largest city in France isELS""", """score""": 2.2E-05, """token""": 16_416, """token_str""": """ELS"""}, ] , ) __UpperCAmelCase : str = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ {"""sequence""": """My name is Patrick""", """score""": 2.1E-05, """token""": 3_499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Te""", """score""": 2E-05, """token""": 2_941, """token_str""": """ Te"""}, {"""sequence""": """My name is Clara""", """score""": 2E-05, """token""": 13_606, """token_str""": """ Clara"""}, ] , ) __UpperCAmelCase : Optional[int] = unmasker("""My name is <mask> <mask>""" , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=6 ) , [ [ { """score""": 2.2E-05, """token""": 35_676, """token_str""": """ Maul""", """sequence""": """<s>My name is Maul<mask></s>""", }, {"""score""": 2.2E-05, """token""": 16_416, """token_str""": """ELS""", """sequence""": """<s>My name isELS<mask></s>"""}, ], [ { """score""": 2.2E-05, """token""": 35_676, """token_str""": """ Maul""", """sequence""": """<s>My name is<mask> Maul</s>""", }, {"""score""": 2.2E-05, """token""": 16_416, """token_str""": """ELS""", """sequence""": """<s>My name is<mask>ELS</s>"""}, ], ] , ) @require_torch_gpu def __A ( self ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : List[str] = pipeline("""fill-mask""" , model="""hf-internal-testing/tiny-random-distilbert""" , device=0 , framework="""pt""" ) # convert model to fp16 pipe.model.half() __UpperCAmelCase : str = pipe("""Paris is the [MASK] of France.""" ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) @slow @require_torch def __A ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase : Any = pipeline(task="""fill-mask""" , model="""distilroberta-base""" , top_k=2 , framework="""pt""" ) self.run_large_test(__UpperCAmelCase ) @slow @require_tf def __A ( self ) -> int: '''simple docstring''' __UpperCAmelCase : int = pipeline(task="""fill-mask""" , model="""distilroberta-base""" , top_k=2 , framework="""tf""" ) self.run_large_test(__UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase ) , [ {"""sequence""": """My name is John""", """score""": 0.008, """token""": 610, """token_str""": """ John"""}, {"""sequence""": """My name is Chris""", """score""": 0.007, """token""": 1_573, """token_str""": """ Chris"""}, ] , ) __UpperCAmelCase : Optional[int] = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(__UpperCAmelCase ) , [ { """sequence""": """The largest city in France is Paris""", """score""": 0.251, """token""": 2_201, """token_str""": """ Paris""", }, { """sequence""": """The largest city in France is Lyon""", """score""": 0.214, """token""": 12_790, """token_str""": """ Lyon""", }, ] , ) __UpperCAmelCase : Optional[int] = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(__UpperCAmelCase ) , [ {"""sequence""": """My name is Patrick""", """score""": 0.005, """token""": 3_499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Clara""", """score""": 0.000, """token""": 13_606, """token_str""": """ Clara"""}, {"""sequence""": """My name is Te""", """score""": 0.000, """token""": 2_941, """token_str""": """ Te"""}, ] , ) @require_torch def __A ( self ) -> List[str]: '''simple docstring''' __UpperCAmelCase : Dict = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , framework="""pt""" ) __UpperCAmelCase : Tuple = None __UpperCAmelCase : int = None self.run_pipeline_test(__UpperCAmelCase , [] ) @require_tf def __A ( self ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase : Dict = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , framework="""tf""" ) __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : str = None self.run_pipeline_test(__UpperCAmelCase , [] ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: '''simple docstring''' if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("""The provided tokenizer has no mask token, (probably reformer or wav2vec2)""" ) __UpperCAmelCase : str = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) __UpperCAmelCase : int = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : Optional[int] = fill_masker.tokenizer __UpperCAmelCase : Union[str, Any] = fill_masker.model __UpperCAmelCase : Tuple = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) __UpperCAmelCase : int = fill_masker([f'This is a {tokenizer.mask_token}'] ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) __UpperCAmelCase : Union[str, Any] = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.'] ) self.assertEqual( __UpperCAmelCase , [ [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ], [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ], ] , ) with self.assertRaises(__UpperCAmelCase ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(__UpperCAmelCase ): fill_masker("""This is""" ) self.run_test_top_k(__UpperCAmelCase , __UpperCAmelCase ) self.run_test_targets(__UpperCAmelCase , __UpperCAmelCase ) self.run_test_top_k_targets(__UpperCAmelCase , __UpperCAmelCase ) self.fill_mask_with_duplicate_targets_and_top_k(__UpperCAmelCase , __UpperCAmelCase ) self.fill_mask_with_multiple_masks(__UpperCAmelCase , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Any: '''simple docstring''' __UpperCAmelCase : Dict = tokenizer.get_vocab() __UpperCAmelCase : Dict = sorted(vocab.keys() )[:2] # Pipeline argument __UpperCAmelCase : Dict = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , targets=__UpperCAmelCase ) __UpperCAmelCase : List[str] = fill_masker(f'This is a {tokenizer.mask_token}' ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) __UpperCAmelCase : Any = {vocab[el] for el in targets} self.assertEqual({el["""token"""] for el in outputs} , __UpperCAmelCase ) __UpperCAmelCase : int = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["""token_str"""] for el in outputs} , set(__UpperCAmelCase ) ) # Call argument __UpperCAmelCase : List[Any] = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) __UpperCAmelCase : Tuple = fill_masker(f'This is a {tokenizer.mask_token}' , targets=__UpperCAmelCase ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) __UpperCAmelCase : List[Any] = {vocab[el] for el in targets} self.assertEqual({el["""token"""] for el in outputs} , __UpperCAmelCase ) __UpperCAmelCase : List[Any] = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["""token_str"""] for el in outputs} , set(__UpperCAmelCase ) ) # Score equivalence __UpperCAmelCase : Dict = fill_masker(f'This is a {tokenizer.mask_token}' , targets=__UpperCAmelCase ) __UpperCAmelCase : Dict = [top_mask["""token_str"""] for top_mask in outputs] __UpperCAmelCase : str = [top_mask["""score"""] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(__UpperCAmelCase ) == set(__UpperCAmelCase ): __UpperCAmelCase : str = fill_masker(f'This is a {tokenizer.mask_token}' , targets=__UpperCAmelCase ) __UpperCAmelCase : int = [top_mask["""score"""] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(__UpperCAmelCase ) , nested_simplify(__UpperCAmelCase ) ) # Raises with invalid with self.assertRaises(__UpperCAmelCase ): __UpperCAmelCase : Any = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(__UpperCAmelCase ): __UpperCAmelCase : Dict = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""""""] ) with self.assertRaises(__UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = fill_masker(f'This is a {tokenizer.mask_token}' , targets="""""" ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' __UpperCAmelCase : Dict = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , top_k=2 ) __UpperCAmelCase : Optional[int] = fill_masker(f'This is a {tokenizer.mask_token}' ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) __UpperCAmelCase : List[Any] = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) __UpperCAmelCase : int = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ] , ) self.assertEqual(nested_simplify(__UpperCAmelCase ) , nested_simplify(__UpperCAmelCase ) ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' __UpperCAmelCase : int = tokenizer.get_vocab() __UpperCAmelCase : List[Any] = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) # top_k=2, ntargets=3 __UpperCAmelCase : Dict = sorted(vocab.keys() )[:3] __UpperCAmelCase : str = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=__UpperCAmelCase ) # If we use the most probably targets, and filter differently, we should still # have the same results __UpperCAmelCase : Tuple = [el["""token_str"""] for el in sorted(__UpperCAmelCase , key=lambda __UpperCAmelCase : x["score"] , reverse=__UpperCAmelCase )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(__UpperCAmelCase ).issubset(__UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=__UpperCAmelCase ) # They should yield exactly the same result self.assertEqual(nested_simplify(__UpperCAmelCase ) , nested_simplify(__UpperCAmelCase ) ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) __UpperCAmelCase : List[Any] = tokenizer.get_vocab() # String duplicates + id duplicates __UpperCAmelCase : Dict = sorted(vocab.keys() )[:3] __UpperCAmelCase : Dict = [targets[0], targets[1], targets[0], targets[2], targets[1]] __UpperCAmelCase : Optional[int] = fill_masker(f'My name is {tokenizer.mask_token}' , targets=__UpperCAmelCase , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(__UpperCAmelCase ) , 3 ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase : List[str] = FillMaskPipeline(model=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) __UpperCAmelCase : Dict = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ], [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ], [ {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, {"""sequence""": ANY(__UpperCAmelCase ), """score""": ANY(__UpperCAmelCase ), """token""": ANY(__UpperCAmelCase ), """token_str""": ANY(__UpperCAmelCase )}, ], ] , )
16
1
import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline __A : Any = { "n_samples": 64, "horizon": 32, "num_inference_steps": 20, "n_guide_steps": 2, # can set to 0 for faster sampling, does not use value network "scale_grad_by_std": True, "scale": 0.1, "eta": 0.0, "t_grad_cutoff": 2, "device": "cpu", } if __name__ == "__main__": __A : Dict = "hopper-medium-v2" __A : int = gym.make(env_name) __A : Tuple = ValueGuidedRLPipeline.from_pretrained( "bglick13/hopper-medium-v2-value-function-hor32", env=env, ) env.seed(0) __A : List[str] = env.reset() __A : Dict = 0 __A : Optional[int] = 0 __A : Any = 1_000 __A : int = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy __A : Dict = pipeline(obs, planning_horizon=32) # execute action in environment __A , __A , __A , __A : Tuple = env.step(denorm_actions) __A : Optional[Any] = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( F'Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:' F' {total_score}' ) # save observations for rendering rollout.append(next_observation.copy()) __A : List[str] = next_observation except KeyboardInterrupt: pass print(F'Total reward: {total_reward}')
273
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer __A : Dict = logging.get_logger(__name__) __A : Any = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __A : Tuple = { "vocab_file": { "squeezebert/squeezebert-uncased": ( "https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt" ), "squeezebert/squeezebert-mnli": "https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt", "squeezebert/squeezebert-mnli-headless": ( "https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt" ), }, "tokenizer_file": { "squeezebert/squeezebert-uncased": ( "https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json" ), "squeezebert/squeezebert-mnli": ( "https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json" ), "squeezebert/squeezebert-mnli-headless": ( "https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json" ), }, } __A : List[Any] = { "squeezebert/squeezebert-uncased": 512, "squeezebert/squeezebert-mnli": 512, "squeezebert/squeezebert-mnli-headless": 512, } __A : List[Any] = { "squeezebert/squeezebert-uncased": {"do_lower_case": True}, "squeezebert/squeezebert-mnli": {"do_lower_case": True}, "squeezebert/squeezebert-mnli-headless": {"do_lower_case": True}, } class A_ (a_ ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = SqueezeBertTokenizer def __init__( self , _A=None , _A=None , _A=True , _A="[UNK]" , _A="[SEP]" , _A="[PAD]" , _A="[CLS]" , _A="[MASK]" , _A=True , _A=None , **_A , ): '''simple docstring''' super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , tokenize_chinese_chars=_A , strip_accents=_A , **_A , ) UpperCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _A ) != do_lower_case or normalizer_state.get('''strip_accents''' , _A ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _A ) != tokenize_chinese_chars ): UpperCAmelCase = getattr(_A , normalizer_state.pop('''type''' ) ) UpperCAmelCase = do_lower_case UpperCAmelCase = strip_accents UpperCAmelCase = tokenize_chinese_chars UpperCAmelCase = normalizer_class(**_A ) UpperCAmelCase = do_lower_case def _lowercase ( self , _A , _A=None ): '''simple docstring''' UpperCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _A , _A = None ): '''simple docstring''' UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _A , _A = None ): '''simple docstring''' UpperCAmelCase = self._tokenizer.model.save(_A , name=_A ) return tuple(_A )
273
1
'''simple docstring''' import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput lowerCamelCase__ : Optional[Any] = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , *lowerCamelCase__ : str , lowerCamelCase__ : Dict=None , lowerCamelCase__ : Optional[Any]=None , lowerCamelCase__ : str=None , **lowerCamelCase__ : Optional[int] ) ->Optional[int]: '''simple docstring''' super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = eval_examples _UpperCAmelCase : List[str] = post_process_function _UpperCAmelCase : List[Any] = quant_trainer_args _UpperCAmelCase : List[str] = 1_28 # default number of calibration samples def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Optional[Any]=None ) ->Any: '''simple docstring''' if calib_dataset is None and self.calib_dataset is None: raise ValueError("Trainer: calibration requires an calib_dataset." ) _UpperCAmelCase : Dict = calib_dataset if calib_dataset is not None else self.calib_dataset _UpperCAmelCase : Union[str, Any] = self._remove_unused_columns(lowerCamelCase__ , description="Calibration" ) return DataLoader( lowerCamelCase__ , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=lowerCamelCase__ , ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[str]=None ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.train_dataset if calib_dataset is None else calib_dataset _UpperCAmelCase : int = self.get_calib_dataloader(lowerCamelCase__ ) _UpperCAmelCase : str = self.model quant_trainer.configure_model(lowerCamelCase__ , self.quant_trainer_args , calib=lowerCamelCase__ ) model.eval() quant_trainer.enable_calibration(lowerCamelCase__ ) logger.info("***** Running calibration *****" ) logger.info(F""" Num examples = {self.calib_num}""" ) logger.info(F""" Batch size = {calib_dataloader.batch_size}""" ) for step, inputs in enumerate(lowerCamelCase__ ): # Prediction step _UpperCAmelCase : Any = self.prediction_step(lowerCamelCase__ , lowerCamelCase__ , prediction_loss_only=lowerCamelCase__ ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(lowerCamelCase__ , self.quant_trainer_args ) _UpperCAmelCase : Dict = model def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Tuple=None , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , lowerCamelCase__ : str = "eval" ) ->Dict: '''simple docstring''' _UpperCAmelCase : Any = self.eval_dataset if eval_dataset is None else eval_dataset _UpperCAmelCase : Tuple = self.get_eval_dataloader(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _UpperCAmelCase : List[str] = self.compute_metrics _UpperCAmelCase : int = None _UpperCAmelCase : Optional[int] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _UpperCAmelCase : Tuple = eval_loop( lowerCamelCase__ , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCamelCase__ , ) finally: _UpperCAmelCase : Tuple = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: _UpperCAmelCase : str = self.post_process_function(lowerCamelCase__ , lowerCamelCase__ , output.predictions ) _UpperCAmelCase : Optional[int] = self.compute_metrics(lowerCamelCase__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): _UpperCAmelCase : List[Any] = metrics.pop(lowerCamelCase__ ) self.log(lowerCamelCase__ ) else: _UpperCAmelCase : List[str] = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) _UpperCAmelCase : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowerCamelCase__ ) return metrics def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : List[Any]=None , lowerCamelCase__ : str = "test" ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.get_test_dataloader(lowerCamelCase__ ) # Temporarily disable metric computation, we will do it in the loop here. _UpperCAmelCase : Dict = self.compute_metrics _UpperCAmelCase : Any = None _UpperCAmelCase : Tuple = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _UpperCAmelCase : List[Any] = eval_loop( lowerCamelCase__ , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCamelCase__ , ) finally: _UpperCAmelCase : Union[str, Any] = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output _UpperCAmelCase : Optional[Any] = self.post_process_function(lowerCamelCase__ , lowerCamelCase__ , output.predictions , "predict" ) _UpperCAmelCase : Tuple = self.compute_metrics(lowerCamelCase__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): _UpperCAmelCase : str = metrics.pop(lowerCamelCase__ ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Tuple="./" ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.eval_dataset _UpperCAmelCase : Tuple = self.get_eval_dataloader(lowerCamelCase__ ) _UpperCAmelCase : Tuple = next(iter(lowerCamelCase__ ) ) # saving device - to make it consistent _UpperCAmelCase : Dict = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) # convert to tuple _UpperCAmelCase : List[Any] = tuple(v.to(lowerCamelCase__ ) for k, v in batch.items() ) logger.info("Converting model to be onnx compatible" ) from pytorch_quantization.nn import TensorQuantizer _UpperCAmelCase : Dict = True _UpperCAmelCase : str = self.model.to(lowerCamelCase__ ) model.eval() model.float() _UpperCAmelCase : int = model.module if hasattr(lowerCamelCase__ , "module" ) else model quant_trainer.configure_model(lowerCamelCase__ , self.quant_trainer_args ) _UpperCAmelCase : Optional[int] = os.path.join(lowerCamelCase__ , "model.onnx" ) logger.info(F"""exporting model to {output_model_file}""" ) _UpperCAmelCase : List[Any] = {0: "batch_size", 1: "seq_len"} torch.onnx.export( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , export_params=lowerCamelCase__ , opset_version=13 , do_constant_folding=lowerCamelCase__ , input_names=["input_ids", "attention_mask", "token_type_ids"] , output_names=["output_start_logits", "output_end_logits"] , dynamic_axes={ "input_ids": axes, "attention_mask": axes, "token_type_ids": axes, "output_start_logits": axes, "output_end_logits": axes, } , verbose=lowerCamelCase__ , ) logger.info("onnx export finished" )
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
"""simple docstring""" import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase = '''▁''' __UpperCamelCase = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : str = BigBirdTokenizer SCREAMING_SNAKE_CASE_ : Any = BigBirdTokenizerFast SCREAMING_SNAKE_CASE_ : List[Any] = True SCREAMING_SNAKE_CASE_ : int = True def __A ( self ) -> List[str]: super().setUp() SCREAMING_SNAKE_CASE = self.tokenizer_class(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ) -> List[Any]: SCREAMING_SNAKE_CASE = '<s>' SCREAMING_SNAKE_CASE = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __A ( self ) -> List[str]: SCREAMING_SNAKE_CASE = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '[MASK]' ) self.assertEqual(len(lowerCAmelCase__ ) , 1_004 ) def __A ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def __A ( self ) -> List[str]: if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE = self.get_tokenizer() SCREAMING_SNAKE_CASE = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE = tokenizer.tokenize(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = rust_tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = rust_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE = tokenizer.encode(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = rust_tokenizer.encode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def __A ( self ) -> int: SCREAMING_SNAKE_CASE = BigBirdTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCAmelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [285, 46, 10, 170, 382] , ) SCREAMING_SNAKE_CASE = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) SCREAMING_SNAKE_CASE = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def __A ( self ) -> Tuple: return BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) @slow def __A ( self ) -> Tuple: SCREAMING_SNAKE_CASE = 'Hello World!' SCREAMING_SNAKE_CASE = [65, 18_536, 2_260, 101, 66] self.assertListEqual(lowerCAmelCase__ , self.big_tokenizer.encode(lowerCAmelCase__ ) ) @slow def __A ( self ) -> int: SCREAMING_SNAKE_CASE = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) # fmt: off SCREAMING_SNAKE_CASE = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(lowerCAmelCase__ , self.big_tokenizer.encode(lowerCAmelCase__ ) ) @require_torch @slow def __A ( self ) -> Tuple: import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence SCREAMING_SNAKE_CASE = list(self.big_tokenizer.get_vocab().keys() )[:10] SCREAMING_SNAKE_CASE = ' '.join(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = self.big_tokenizer.encode_plus(lowerCAmelCase__ , return_tensors='pt' , return_token_type_ids=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = BigBirdConfig(attention_type='original_full' ) SCREAMING_SNAKE_CASE = BigBirdModel(lowerCAmelCase__ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowerCAmelCase__ ) model(**lowerCAmelCase__ ) @slow def __A ( self ) -> List[str]: SCREAMING_SNAKE_CASE = BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) SCREAMING_SNAKE_CASE = tokenizer.decode(tokenizer('Paris is the [MASK].' ).input_ids ) self.assertTrue(decoded_text == '[CLS] Paris is the[MASK].[SEP]' ) @slow def __A ( self ) -> Optional[int]: # fmt: off SCREAMING_SNAKE_CASE = {'input_ids': [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name='google/bigbird-roberta-base' , revision='215c99f1600e06f83acce68422f2035b2b5c3510' , )
113
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE__ = {"configuration_focalnet": ["FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FocalNetConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FocalNetForImageClassification", "FocalNetForMaskedImageModeling", "FocalNetBackbone", "FocalNetModel", "FocalNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
46
0
import math A_ :List[str] = 10 A_ :Optional[int] = 7 A_ :int = BALLS_PER_COLOUR * NUM_COLOURS def lowerCAmelCase ( a_ = 20 ) -> Dict: __UpperCamelCase : str =math.comb(a__ ,a__ ) __UpperCamelCase : Dict =math.comb(NUM_BALLS - BALLS_PER_COLOUR ,a__ ) __UpperCamelCase : Any =NUM_COLOURS * (1 - missing_colour / total) return F'{result:.9f}' if __name__ == "__main__": print(solution(20))
369
def A ( a_ ) -> bool: return sum(i for i in range(1 ,number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('''Program to check whether a number is a Perfect number or not...''') A_ :List[str] = int(input('''Enter number: ''').strip()) print(f"{number} is {'' if perfect(number) else 'not '}a Perfect Number.")
245
0
def lowerCamelCase_ ( _a ): """simple docstring""" return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(_a ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__('''doctest''').testmod()
131
from copy import deepcopy class _a : def __init__( self : List[str] , _SCREAMING_SNAKE_CASE : list[int] | None = None , _SCREAMING_SNAKE_CASE : int | None = None )-> None: if arr is None and size is not None: lowerCAmelCase__ : str = size lowerCAmelCase__ : Optional[Any] = [0] * size elif arr is not None: self.init(_SCREAMING_SNAKE_CASE ) else: raise ValueError('''Either arr or size must be specified''' ) def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : list[int] )-> None: lowerCAmelCase__ : int = len(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = deepcopy(_SCREAMING_SNAKE_CASE ) for i in range(1 , self.size ): lowerCAmelCase__ : Optional[Any] = self.next_(_SCREAMING_SNAKE_CASE ) if j < self.size: self.tree[j] += self.tree[i] def UpperCAmelCase__( self : Optional[int] )-> list[int]: lowerCAmelCase__ : Any = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): lowerCAmelCase__ : Dict = self.next_(_SCREAMING_SNAKE_CASE ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def UpperCAmelCase__( _SCREAMING_SNAKE_CASE : int )-> int: return index + (index & (-index)) @staticmethod def UpperCAmelCase__( _SCREAMING_SNAKE_CASE : int )-> int: return index - (index & (-index)) def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int )-> None: if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value lowerCAmelCase__ : List[str] = self.next_(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int )-> None: self.add(_SCREAMING_SNAKE_CASE , value - self.get(_SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase__( self : Any , _SCREAMING_SNAKE_CASE : int )-> int: if right == 0: return 0 lowerCAmelCase__ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] lowerCAmelCase__ : Dict = self.prev(_SCREAMING_SNAKE_CASE ) return result def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int )-> int: return self.prefix(_SCREAMING_SNAKE_CASE ) - self.prefix(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : int )-> int: return self.query(_SCREAMING_SNAKE_CASE , index + 1 ) def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : int )-> int: value -= self.tree[0] if value < 0: return -1 lowerCAmelCase__ : Any = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 lowerCAmelCase__ : List[Any] = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
131
1
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" _snake_case = AlbertConfig.from_json_file(_UpperCamelCase ) print(F"""Building PyTorch model from configuration: {config}""" ) _snake_case = AlbertForPreTraining(_UpperCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_albert(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _UpperCamelCase ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __A = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
278
from ..utils import DummyObject, requires_backends class lowercase_ ( metaclass=__lowercase ): UpperCamelCase_ : Optional[int] = ["speech"] def __init__( self : str , *A__ : List[str] , **A__ : Tuple ) -> Optional[Any]: requires_backends(self , ['''speech'''] ) class lowercase_ ( metaclass=__lowercase ): UpperCamelCase_ : Optional[Any] = ["speech"] def __init__( self : Dict , *A__ : int , **A__ : int ) -> Tuple: requires_backends(self , ['''speech'''] )
278
1
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis_float32 (there's also the fix_lavis branch) # also note: to convert Vicuna checkpoints, we had to include /home/niels/python_projects/checkpoints/FastChat/vicuna-7b in lavis/configs/models/blip2/blip2_instruct_vicuna7b.yaml # same for Vicuna-13b from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipImageProcessor, InstructBlipConfig, InstructBlipForConditionalGeneration, InstructBlipProcessor, InstructBlipQFormerConfig, InstructBlipVisionConfig, LlamaConfig, LlamaTokenizerFast, TaConfig, TaTokenizerFast, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowercase__ ( ) -> Optional[int]: """simple docstring""" __UpperCamelCase = 'https://raw.githubusercontent.com/salesforce/LAVIS/main/docs/_static/Confusing-Pictures.jpg' __UpperCamelCase = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ).convert('RGB' ) return image def lowercase__ ( __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [] # fmt: off # vision encoder rename_keys.append(('visual_encoder.cls_token', 'vision_model.embeddings.class_embedding') ) rename_keys.append(('visual_encoder.pos_embed', 'vision_model.embeddings.position_embedding') ) rename_keys.append(('visual_encoder.patch_embed.proj.weight', 'vision_model.embeddings.patch_embedding.weight') ) rename_keys.append(('visual_encoder.patch_embed.proj.bias', 'vision_model.embeddings.patch_embedding.bias') ) rename_keys.append(('ln_vision.weight', 'vision_model.post_layernorm.weight') ) rename_keys.append(('ln_vision.bias', 'vision_model.post_layernorm.bias') ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.weight''', F'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.bias''', F'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.weight''', F'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.bias''', F'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.qkv.weight''', F'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.weight''', F'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.bias''', F'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(('Qformer.bert.embeddings.LayerNorm.weight', 'qformer.embeddings.layernorm.weight') ) rename_keys.append(('Qformer.bert.embeddings.LayerNorm.bias', 'qformer.embeddings.layernorm.bias') ) # fmt: on return rename_keys def lowercase__ ( __lowercase : List[Any] , __lowercase : Any , __lowercase : Any ) -> List[str]: """simple docstring""" __UpperCamelCase = dct.pop(__lowercase ) __UpperCamelCase = val def lowercase__ ( __lowercase : str , __lowercase : Tuple ) -> Union[str, Any]: """simple docstring""" for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __UpperCamelCase = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.q_bias''' ) __UpperCamelCase = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict __UpperCamelCase = torch.cat((q_bias, torch.zeros_like(__lowercase , requires_grad=__lowercase ), v_bias) ) __UpperCamelCase = qkv_bias def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = 364 if 'coco' in model_name else 224 __UpperCamelCase = InstructBlipVisionConfig(image_size=__lowercase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "t5-xl" in model_name: __UpperCamelCase = TaConfig.from_pretrained('google/flan-t5-xl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __UpperCamelCase = TaConfig.from_pretrained('google/flan-t5-xxl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() elif "vicuna-7b" in model_name: __UpperCamelCase = LlamaConfig.from_pretrained('decapoda-research/llama-7b-hf' , vocab_size=32001 ).to_dict() elif "vicuna-13b" in model_name: __UpperCamelCase = LlamaConfig.from_pretrained('decapoda-research/llama-13b-hf' , vocab_size=32001 ).to_dict() else: raise ValueError('Model name not supported' ) # the authors add one special "[DEC]" token to the vocab of Q-Former, hence vocab size = 30522 + 1 __UpperCamelCase = InstructBlipQFormerConfig(vocab_size=30523 ).to_dict() __UpperCamelCase = InstructBlipConfig(vision_config=__lowercase , text_config=__lowercase , qformer_config=__lowercase ) return config, image_size @torch.no_grad() def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict=None , __lowercase : Dict=False ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = AutoTokenizer.from_pretrained('bert-base-uncased' , truncation_side='left' ) qformer_tokenizer.add_special_tokens({'bos_token': '[DEC]'} ) if "t5" in model_name: __UpperCamelCase = TaTokenizerFast.from_pretrained('google/flan-t5-xl' , truncation_side='left' ) elif "vicuna" in model_name: # the following was used in the original implementation: # tokenizer = LlamaTokenizer.from_pretrained("huggyllama/llama-7b", use_fast=False, truncation_side="left") # tokenizer.add_special_tokens({"pad_token": "[PAD]"}) # tokenizer.add_special_tokens({"bos_token": "</s>"}) # tokenizer.add_special_tokens({"eos_token": "</s>"}) # tokenizer.add_special_tokens({"unk_token": "</s>"}) __UpperCamelCase = LlamaTokenizerFast.from_pretrained( 'huggyllama/llama-7b' , truncation_side='left' , bos_token='</s>' , unk_token='</s>' ) tokenizer.add_special_tokens({'pad_token': '[PAD]'} ) __UpperCamelCase , __UpperCamelCase = get_blipa_config(__lowercase ) __UpperCamelCase = InstructBlipForConditionalGeneration(__lowercase ).eval() __UpperCamelCase = { 'instructblip-vicuna-7b': ('blip2_vicuna_instruct', 'vicuna7b'), 'instructblip-vicuna-13b': ('blip2_vicuna_instruct', 'vicuna13b'), 'instructblip-flan-t5-xl': ('blip2_t5_instruct', 'flant5xl'), 'instructblip-flan-t5-xxl': ('blip2_t5_instruct', 'flant5xxl'), } __UpperCamelCase , __UpperCamelCase = model_name_to_original[model_name] # load original model print('Loading original model...' ) __UpperCamelCase = 'cuda:1' if torch.cuda.is_available() else 'cpu' __UpperCamelCase = 'cuda:2' if torch.cuda.is_available() else 'cpu' __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = load_model_and_preprocess( name=__lowercase , model_type=__lowercase , is_eval=__lowercase , device=__lowercase ) original_model.eval() print('Done!' ) # update state dict keys __UpperCamelCase = original_model.state_dict() __UpperCamelCase = create_rename_keys(__lowercase ) for src, dest in rename_keys: rename_key(__lowercase , __lowercase , __lowercase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __UpperCamelCase = state_dict.pop(__lowercase ) if key.startswith('Qformer.bert' ): __UpperCamelCase = key.replace('Qformer.bert' , 'qformer' ) if "attention.self" in key: __UpperCamelCase = key.replace('self' , 'attention' ) if "llm_proj" in key: __UpperCamelCase = key.replace('llm_proj' , 'language_projection' ) if "t5_proj" in key: __UpperCamelCase = key.replace('t5_proj' , 'language_projection' ) if key.startswith('llm_model' ): __UpperCamelCase = key.replace('llm_model' , 'language_model' ) if key.startswith('t5' ): __UpperCamelCase = key.replace('t5' , 'language' ) __UpperCamelCase = val # read in qv biases read_in_q_v_bias(__lowercase , __lowercase ) # note: weights get loaded in torch.float32 by default hf_model.load_state_dict(__lowercase , strict=__lowercase ) __UpperCamelCase = load_demo_image() __UpperCamelCase = 'What is unusual about this image?' # create processor __UpperCamelCase = BlipImageProcessor( size={'height': image_size, 'width': image_size} , image_mean=__lowercase , image_std=__lowercase ) __UpperCamelCase = InstructBlipProcessor( image_processor=__lowercase , tokenizer=__lowercase , qformer_tokenizer=__lowercase , ) __UpperCamelCase = processor(images=__lowercase , text=__lowercase , return_tensors='pt' ).to(__lowercase ) # make sure processor creates exact same pixel values __UpperCamelCase = vis_processors['eval'](__lowercase ).unsqueeze(0 ).to(__lowercase ) __UpperCamelCase = inputs.pixel_values assert torch.allclose(original_pixel_values.to(pixel_values.device ) , __lowercase ) original_model.to(__lowercase ) hf_model.to(__lowercase ) with torch.no_grad(): if "vicuna" in model_name: __UpperCamelCase = original_model({'image': original_pixel_values, 'text_input': [prompt]} ).logits __UpperCamelCase = hf_model(**__lowercase ).logits else: __UpperCamelCase = original_model( {'image': original_pixel_values, 'text_input': [prompt], 'text_output': ['\n']} ).logits __UpperCamelCase = tokenizer('\n' , return_tensors='pt' ).input_ids.to(__lowercase ) __UpperCamelCase = label_input_ids.masked_fill(label_input_ids == tokenizer.pad_token_id , -100 ) __UpperCamelCase = hf_model(**__lowercase , labels=__lowercase ).logits print('First values of original logits:' , original_logits[0, :3, :3] ) print('First values of HF logits:' , logits[0, :3, :3] ) # assert values assert original_logits.shape == logits.shape __UpperCamelCase = 1e-4 if 'vicuna' in model_name else 1e-5 assert torch.allclose(original_logits.to(logits.device ) , __lowercase , atol=__lowercase ) print('Looks ok!' ) print('Generating with original model...' ) __UpperCamelCase = original_model.generate({'image': original_pixel_values, 'prompt': prompt} , num_beams=5 ) # important: we need to cast the weights of the HF model to the appropriate type print('Generating with HF model...' ) __UpperCamelCase = hf_model.generate( **__lowercase , do_sample=__lowercase , num_beams=5 , max_length=256 , min_length=1 , top_p=0.9 , repetition_penalty=1.5 , length_penalty=1.0 , temperature=1 , ) if "vicuna" in model_name: # convert output id 0 to 2 (eos_token_id) # TODO add this in the generate method? __UpperCamelCase = 2 print('Original generation:' , __lowercase ) __UpperCamelCase = processor.batch_decode(__lowercase , skip_special_tokens=__lowercase ) __UpperCamelCase = [text.strip() for text in output_text] print('HF generation:' , __lowercase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(__lowercase ) hf_model.save_pretrained(__lowercase ) if push_to_hub: processor.push_to_hub(F'''Salesforce/{model_name}''' ) hf_model.push_to_hub(F'''Salesforce/{model_name}''' ) if __name__ == "__main__": a__ : str =argparse.ArgumentParser() a__ : List[str] =[ '''instructblip-vicuna-7b''', '''instructblip-vicuna-13b''', '''instructblip-flan-t5-xl''', '''instructblip-flan-t5-xxl''', ] parser.add_argument( '''--model_name''', default='''instructblip-flan-t5-xl''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) a__ : Any =parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
53
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
1
"""simple docstring""" import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values _a = argparse.ArgumentParser() parser.add_argument('--user', type=str, default='ubuntu') parser.add_argument('--host', type=str, default='localhost') parser.add_argument('--key_path', type=str, default=None) parser.add_argument('--instance', type=str, default='V100:1') parser.add_argument('--provider', type=str, default='cheapest') parser.add_argument('--use_spot', type=bool, default=False) parser.add_argument('--example', type=str, default='pytorch/text-generation/run_generation.py') _a , _a = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('Cannot specify both BYO and on-demand cluster args') _a = rh.cluster( name='rh-cluster', ips=[args.host], ssh_creds={'ssh_user': args.user, 'ssh_private_key': args.key_path} ) else: _a = rh.cluster( name='rh-cluster', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) _a = args.example.rsplit('/', 1)[0] # Set up remote environment cluster.install_packages(['pip:./']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F"pip install -r transformers/examples/{example_dir}/requirements.txt"]) cluster.run(['pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F"python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}"]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
144
"""simple docstring""" import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset _a = 'bert-base-cased' _a = 'google/pegasus-xsum' _a = [' Sam ate lunch today.', 'Sams lunch ingredients.'] _a = ['A very interesting story about what I ate for lunch.', 'Avocado, celery, turkey, coffee'] _a = 'patrickvonplaten/t5-tiny-random' _a = 'sshleifer/bart-tiny-random' _a = 'sshleifer/tiny-mbart' _a = 'sshleifer/tiny-marian-en-de' def _A ( UpperCamelCase_ : Path, UpperCamelCase_ : list) -> Optional[Any]: '''simple docstring''' __lowercase = "\n".join(UpperCamelCase_) Path(UpperCamelCase_).open("w").writelines(UpperCamelCase_) def _A ( UpperCamelCase_ : Union[str, Any]) -> int: '''simple docstring''' for split in ["train", "val", "test"]: _dump_articles(os.path.join(UpperCamelCase_, F"""{split}.source"""), UpperCamelCase_) _dump_articles(os.path.join(UpperCamelCase_, F"""{split}.target"""), UpperCamelCase_) return tmp_dir class _lowerCAmelCase ( lowercase ): """simple docstring""" @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ], ) @slow def _lowercase ( self : List[Any], UpperCAmelCase__ : Optional[Any] ): __lowercase = AutoTokenizer.from_pretrained(UpperCAmelCase__ ) __lowercase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __lowercase = max(len(tokenizer.encode(UpperCAmelCase__ ) ) for a in ARTICLES ) __lowercase = max(len(tokenizer.encode(UpperCAmelCase__ ) ) for a in SUMMARIES ) __lowercase = 4 __lowercase = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated __lowercase ,__lowercase = "ro_RO", "de_DE" # ignored for all but mbart, but never causes error. __lowercase = SeqaSeqDataset( UpperCAmelCase__, data_dir=UpperCAmelCase__, type_path="train", max_source_length=UpperCAmelCase__, max_target_length=UpperCAmelCase__, src_lang=UpperCAmelCase__, tgt_lang=UpperCAmelCase__, ) __lowercase = DataLoader(UpperCAmelCase__, batch_size=2, collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(UpperCAmelCase__, UpperCAmelCase__ ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place __lowercase = shift_tokens_right(batch["labels"], tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def _lowercase ( self : Dict, UpperCAmelCase__ : int ): __lowercase = AutoTokenizer.from_pretrained(UpperCAmelCase__ ) __lowercase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __lowercase = max(len(tokenizer.encode(UpperCAmelCase__ ) ) for a in ARTICLES ) __lowercase = max(len(tokenizer.encode(UpperCAmelCase__ ) ) for a in SUMMARIES ) __lowercase = 4 __lowercase = LegacySeqaSeqDataset( UpperCAmelCase__, data_dir=UpperCAmelCase__, type_path="train", max_source_length=2_0, max_target_length=UpperCAmelCase__, ) __lowercase = DataLoader(UpperCAmelCase__, batch_size=2, collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 2_0 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def _lowercase ( self : str ): __lowercase = AutoTokenizer.from_pretrained("facebook/mbart-large-cc25" ) __lowercase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) __lowercase = tmp_dir.joinpath("train.source" ).open().readlines() __lowercase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(UpperCAmelCase__, UpperCAmelCase__, 1_2_8, UpperCAmelCase__ ) __lowercase = {x.name for x in tmp_dir.iterdir()} __lowercase = {x.name for x in save_dir.iterdir()} __lowercase = save_dir.joinpath("train.source" ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(UpperCAmelCase__ ) < len(UpperCAmelCase__ ) assert len(UpperCAmelCase__ ) == 1 assert len(packed_examples[0] ) == sum(len(UpperCAmelCase__ ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE, reason="This test requires fairseq" ) def _lowercase ( self : List[str] ): if not FAIRSEQ_AVAILABLE: return __lowercase ,__lowercase ,__lowercase = self._get_dataset(max_len=6_4 ) __lowercase = 6_4 __lowercase = ds.make_dynamic_sampler(UpperCAmelCase__, required_batch_size_multiple=UpperCAmelCase__ ) __lowercase = [len(UpperCAmelCase__ ) for x in batch_sampler] assert len(set(UpperCAmelCase__ ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) # no dropped or added examples __lowercase = DataLoader(UpperCAmelCase__, batch_sampler=UpperCAmelCase__, collate_fn=ds.collate_fn, num_workers=2 ) __lowercase = [] __lowercase = [] for batch in data_loader: __lowercase = batch["input_ids"].shape __lowercase = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple __lowercase = np.product(batch["input_ids"].shape ) num_src_per_batch.append(UpperCAmelCase__ ) if num_src_tokens > (max_tokens * 1.1): failures.append(UpperCAmelCase__ ) assert num_src_per_batch[0] == max(UpperCAmelCase__ ) if failures: raise AssertionError(F"""too many tokens in {len(UpperCAmelCase__ )} batches""" ) def _lowercase ( self : Any ): __lowercase ,__lowercase ,__lowercase = self._get_dataset(max_len=5_1_2 ) __lowercase = 2 __lowercase = ds.make_sortish_sampler(UpperCAmelCase__, shuffle=UpperCAmelCase__ ) __lowercase = DataLoader(UpperCAmelCase__, batch_size=UpperCAmelCase__, collate_fn=ds.collate_fn, num_workers=2 ) __lowercase = DataLoader(UpperCAmelCase__, batch_size=UpperCAmelCase__, collate_fn=ds.collate_fn, num_workers=2, sampler=UpperCAmelCase__ ) __lowercase = tokenizer.pad_token_id def count_pad_tokens(UpperCAmelCase__ : Any, UpperCAmelCase__ : Union[str, Any]="input_ids" ): return [batch[k].eq(UpperCAmelCase__ ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(UpperCAmelCase__, k="labels" ) ) < sum(count_pad_tokens(UpperCAmelCase__, k="labels" ) ) assert sum(count_pad_tokens(UpperCAmelCase__ ) ) < sum(count_pad_tokens(UpperCAmelCase__ ) ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) def _lowercase ( self : List[Any], UpperCAmelCase__ : int=1_0_0_0, UpperCAmelCase__ : str=1_2_8 ): if os.getenv("USE_REAL_DATA", UpperCAmelCase__ ): __lowercase = "examples/seq2seq/wmt_en_ro" __lowercase = max_len * 2 * 6_4 if not Path(UpperCAmelCase__ ).joinpath("train.len" ).exists(): save_len_file(UpperCAmelCase__, UpperCAmelCase__ ) else: __lowercase = "examples/seq2seq/test_data/wmt_en_ro" __lowercase = max_len * 4 save_len_file(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = AutoTokenizer.from_pretrained(UpperCAmelCase__ ) __lowercase = SeqaSeqDataset( UpperCAmelCase__, data_dir=UpperCAmelCase__, type_path="train", max_source_length=UpperCAmelCase__, max_target_length=UpperCAmelCase__, n_obs=UpperCAmelCase__, ) return ds, max_tokens, tokenizer def _lowercase ( self : Union[str, Any] ): __lowercase ,__lowercase ,__lowercase = self._get_dataset() __lowercase = set(DistributedSortishSampler(UpperCAmelCase__, 2_5_6, num_replicas=2, rank=0, add_extra_examples=UpperCAmelCase__ ) ) __lowercase = set(DistributedSortishSampler(UpperCAmelCase__, 2_5_6, num_replicas=2, rank=1, add_extra_examples=UpperCAmelCase__ ) ) assert idsa.intersection(UpperCAmelCase__ ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ], ) def _lowercase ( self : int, UpperCAmelCase__ : Optional[Any] ): __lowercase = AutoTokenizer.from_pretrained(UpperCAmelCase__, use_fast=UpperCAmelCase__ ) if tok_name == MBART_TINY: __lowercase = SeqaSeqDataset( UpperCAmelCase__, data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ), type_path="train", max_source_length=4, max_target_length=8, src_lang="EN", tgt_lang="FR", ) __lowercase = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: __lowercase = SeqaSeqDataset( UpperCAmelCase__, data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ), type_path="train", max_source_length=4, max_target_length=8, ) __lowercase = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(UpperCAmelCase__ ) == 1 if tok_name == BART_TINY else len(UpperCAmelCase__ ) == 0
144
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class lowercase__ ( _UpperCAmelCase ): a_ ='''gpt_bigcode''' a_ =['''past_key_values'''] a_ ={ '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self , __UpperCAmelCase=50257 , __UpperCAmelCase=1024 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=None , __UpperCAmelCase="gelu_pytorch_tanh" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=50256 , __UpperCAmelCase=50256 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , )-> Any: '''simple docstring''' lowerCAmelCase__ = vocab_size lowerCAmelCase__ = n_positions lowerCAmelCase__ = n_embd lowerCAmelCase__ = n_layer lowerCAmelCase__ = n_head lowerCAmelCase__ = n_inner lowerCAmelCase__ = activation_function lowerCAmelCase__ = resid_pdrop lowerCAmelCase__ = embd_pdrop lowerCAmelCase__ = attn_pdrop lowerCAmelCase__ = layer_norm_epsilon lowerCAmelCase__ = initializer_range lowerCAmelCase__ = scale_attn_weights lowerCAmelCase__ = use_cache lowerCAmelCase__ = attention_softmax_in_fpaa lowerCAmelCase__ = scale_attention_softmax_in_fpaa lowerCAmelCase__ = multi_query lowerCAmelCase__ = bos_token_id lowerCAmelCase__ = eos_token_id super().__init__(bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase )
340
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: A : Tuple = None A : Optional[Any] = logging.get_logger(__name__) A : Tuple = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} A : List[str] = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } A : List[str] = { '''albert-base-v1''': 5_1_2, '''albert-large-v1''': 5_1_2, '''albert-xlarge-v1''': 5_1_2, '''albert-xxlarge-v1''': 5_1_2, '''albert-base-v2''': 5_1_2, '''albert-large-v2''': 5_1_2, '''albert-xlarge-v2''': 5_1_2, '''albert-xxlarge-v2''': 5_1_2, } A : Optional[int] = '''▁''' class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : str = VOCAB_FILES_NAMES __lowerCamelCase : List[Any] = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase : List[str] = AlbertTokenizer def __init__( self : Tuple , __lowerCAmelCase : Tuple=None , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : str=False , __lowerCAmelCase : Union[str, Any]="[CLS]" , __lowerCAmelCase : int="[SEP]" , __lowerCAmelCase : Dict="<unk>" , __lowerCAmelCase : Dict="[SEP]" , __lowerCAmelCase : Union[str, Any]="<pad>" , __lowerCAmelCase : str="[CLS]" , __lowerCAmelCase : int="[MASK]" , **__lowerCAmelCase : Optional[Any] , ) -> Optional[Any]: """simple docstring""" A__ = ( AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase , normalized=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else mask_token ) super().__init__( __lowerCAmelCase , tokenizer_file=__lowerCAmelCase , do_lower_case=__lowerCAmelCase , remove_space=__lowerCAmelCase , keep_accents=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , cls_token=__lowerCAmelCase , mask_token=__lowerCAmelCase , **__lowerCAmelCase , ) A__ = do_lower_case A__ = remove_space A__ = keep_accents A__ = vocab_file A__ = False if not self.vocab_file else True def a_ ( self : List[Any] , __lowerCAmelCase : List[int] , __lowerCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def a_ ( self : Union[str, Any] , __lowerCAmelCase : List[int] , __lowerCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a_ ( self : Tuple , __lowerCAmelCase : str , __lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(__lowerCAmelCase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return A__ = os.path.join( __lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCAmelCase ): copyfile(self.vocab_file , __lowerCAmelCase ) return (out_vocab_file,)
274
0
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = MODEL_FOR_CAUSAL_LM_MAPPING lowerCAmelCase__ = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def A__ ( self ) -> List[Any]: '''simple docstring''' lowercase_ = pipeline(task="text-generation" , model="sshleifer/tiny-ctrl" , framework="pt" ) # Using `do_sample=False` to force deterministic output lowercase_ = text_generator("This is a test" , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ { "generated_text": ( "This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope." " oscope. FiliFili@@" ) } ] , ) lowercase_ = text_generator(["This is a test", "This is a second test"] ) self.assertEqual( UpperCAmelCase , [ [ { "generated_text": ( "This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope." " oscope. FiliFili@@" ) } ], [ { "generated_text": ( "This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy" " oscope. oscope. FiliFili@@" ) } ], ] , ) lowercase_ = text_generator("This is a test" , do_sample=UpperCAmelCase , num_return_sequences=2 , return_tensors=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ {"generated_token_ids": ANY(UpperCAmelCase )}, {"generated_token_ids": ANY(UpperCAmelCase )}, ] , ) lowercase_ = text_generator.model.config.eos_token_id lowercase_ = "<pad>" lowercase_ = text_generator( ["This is a test", "This is a second test"] , do_sample=UpperCAmelCase , num_return_sequences=2 , batch_size=2 , return_tensors=UpperCAmelCase , ) self.assertEqual( UpperCAmelCase , [ [ {"generated_token_ids": ANY(UpperCAmelCase )}, {"generated_token_ids": ANY(UpperCAmelCase )}, ], [ {"generated_token_ids": ANY(UpperCAmelCase )}, {"generated_token_ids": ANY(UpperCAmelCase )}, ], ] , ) @require_tf def A__ ( self ) -> Tuple: '''simple docstring''' lowercase_ = pipeline(task="text-generation" , model="sshleifer/tiny-ctrl" , framework="tf" ) # Using `do_sample=False` to force deterministic output lowercase_ = text_generator("This is a test" , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ { "generated_text": ( "This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵" " please," ) } ] , ) lowercase_ = text_generator(["This is a test", "This is a second test"] , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ [ { "generated_text": ( "This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵" " please," ) } ], [ { "generated_text": ( "This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes" " Cannes 閲閲Cannes Cannes Cannes 攵 please," ) } ], ] , ) def A__ ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: '''simple docstring''' lowercase_ = TextGenerationPipeline(model=UpperCAmelCase , tokenizer=UpperCAmelCase ) return text_generator, ["This is a test", "Another test"] def A__ ( self ) -> List[Any]: '''simple docstring''' lowercase_ = "Hello I believe in" lowercase_ = pipeline("text-generation" , model="hf-internal-testing/tiny-random-gpt2" ) lowercase_ = text_generator(UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [{"generated_text": "Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe"}] , ) lowercase_ = text_generator(UpperCAmelCase , stop_sequence=" fe" ) self.assertEqual(UpperCAmelCase , [{"generated_text": "Hello I believe in fe"}] ) def A__ ( self , UpperCAmelCase , UpperCAmelCase ) -> Dict: '''simple docstring''' lowercase_ = text_generator.model lowercase_ = text_generator.tokenizer lowercase_ = text_generator("This is a test" ) self.assertEqual(UpperCAmelCase , [{"generated_text": ANY(UpperCAmelCase )}] ) self.assertTrue(outputs[0]["generated_text"].startswith("This is a test" ) ) lowercase_ = text_generator("This is a test" , return_full_text=UpperCAmelCase ) self.assertEqual(UpperCAmelCase , [{"generated_text": ANY(UpperCAmelCase )}] ) self.assertNotIn("This is a test" , outputs[0]["generated_text"] ) lowercase_ = pipeline(task="text-generation" , model=UpperCAmelCase , tokenizer=UpperCAmelCase , return_full_text=UpperCAmelCase ) lowercase_ = text_generator("This is a test" ) self.assertEqual(UpperCAmelCase , [{"generated_text": ANY(UpperCAmelCase )}] ) self.assertNotIn("This is a test" , outputs[0]["generated_text"] ) lowercase_ = text_generator("This is a test" , return_full_text=UpperCAmelCase ) self.assertEqual(UpperCAmelCase , [{"generated_text": ANY(UpperCAmelCase )}] ) self.assertTrue(outputs[0]["generated_text"].startswith("This is a test" ) ) lowercase_ = text_generator(["This is great !", "Something else"] , num_return_sequences=2 , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ [{"generated_text": ANY(UpperCAmelCase )}, {"generated_text": ANY(UpperCAmelCase )}], [{"generated_text": ANY(UpperCAmelCase )}, {"generated_text": ANY(UpperCAmelCase )}], ] , ) if text_generator.tokenizer.pad_token is not None: lowercase_ = text_generator( ["This is great !", "Something else"] , num_return_sequences=2 , batch_size=2 , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ [{"generated_text": ANY(UpperCAmelCase )}, {"generated_text": ANY(UpperCAmelCase )}], [{"generated_text": ANY(UpperCAmelCase )}, {"generated_text": ANY(UpperCAmelCase )}], ] , ) with self.assertRaises(UpperCAmelCase ): lowercase_ = text_generator("test" , return_full_text=UpperCAmelCase , return_text=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): lowercase_ = text_generator("test" , return_full_text=UpperCAmelCase , return_tensors=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): lowercase_ = text_generator("test" , return_text=UpperCAmelCase , return_tensors=UpperCAmelCase ) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): lowercase_ = text_generator("" ) self.assertEqual(UpperCAmelCase , [{"generated_text": ANY(UpperCAmelCase )}] ) else: with self.assertRaises((ValueError, AssertionError) ): lowercase_ = text_generator("" ) if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. lowercase_ = ["RwkvForCausalLM", "XGLMForCausalLM", "GPTNeoXForCausalLM"] if ( tokenizer.model_max_length < 10000 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError) ): text_generator("This is a test" * 500 , max_new_tokens=20 ) lowercase_ = text_generator("This is a test" * 500 , handle_long_generation="hole" , max_new_tokens=20 ) # Hole strategy cannot work with self.assertRaises(UpperCAmelCase ): text_generator( "This is a test" * 500 , handle_long_generation="hole" , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def A__ ( self ) -> Tuple: '''simple docstring''' import torch # Classic `model_kwargs` lowercase_ = pipeline( model="hf-internal-testing/tiny-random-bloom" , model_kwargs={"device_map": "auto", "torch_dtype": torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowercase_ = pipe("This is a test" ) self.assertEqual( UpperCAmelCase , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) lowercase_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" , torch_dtype=torch.bfloataa ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowercase_ = pipe("This is a test" ) self.assertEqual( UpperCAmelCase , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 lowercase_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa ) lowercase_ = pipe("This is a test" ) self.assertEqual( UpperCAmelCase , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) @require_torch @require_torch_gpu def A__ ( self ) -> Optional[Any]: '''simple docstring''' import torch lowercase_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device=0 , torch_dtype=torch.floataa ) pipe("This is a test" ) @require_torch @require_accelerate @require_torch_gpu def A__ ( self ) -> Tuple: '''simple docstring''' import torch lowercase_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" , torch_dtype=torch.floataa ) pipe("This is a test" , do_sample=UpperCAmelCase , top_p=0.5 ) def A__ ( self ) -> Tuple: '''simple docstring''' lowercase_ = "Hello world" lowercase_ = pipeline("text-generation" , model="hf-internal-testing/tiny-random-gpt2" ) if text_generator.model.framework == "tf": lowercase_ = logging.get_logger("transformers.generation.tf_utils" ) else: lowercase_ = logging.get_logger("transformers.generation.utils" ) lowercase_ = "Both `max_new_tokens`" # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(UpperCAmelCase ) as cl: lowercase_ = text_generator(UpperCAmelCase , max_length=10 , max_new_tokens=1 ) self.assertIn(UpperCAmelCase , cl.out ) # The user only sets one -> no warning with CaptureLogger(UpperCAmelCase ) as cl: lowercase_ = text_generator(UpperCAmelCase , max_new_tokens=1 ) self.assertNotIn(UpperCAmelCase , cl.out ) with CaptureLogger(UpperCAmelCase ) as cl: lowercase_ = text_generator(UpperCAmelCase , max_length=10 ) self.assertNotIn(UpperCAmelCase , cl.out )
297
def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: int ): '''simple docstring''' return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("""Program to check whether a number is a Perfect number or not...""") SCREAMING_SNAKE_CASE__ = int(input("""Enter number: """).strip()) print(f"""{number} is {'' if perfect(number) else 'not '}a Perfect Number.""")
297
1
from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer import PriorTransformer from .ta_film_transformer import TaFilmDecoder from .transformer_ad import TransformeraDModel from .unet_ad import UNetaDModel from .unet_ad import UNetaDModel from .unet_ad_condition import UNetaDConditionModel from .unet_ad_condition import UNetaDConditionModel from .vq_model import VQModel if is_flax_available(): from .controlnet_flax import FlaxControlNetModel from .unet_ad_condition_flax import FlaxUNetaDConditionModel from .vae_flax import FlaxAutoencoderKL
51
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCAmelCase : str = { 'configuration_funnel': ['FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FunnelConfig'], 'convert_funnel_original_tf_checkpoint_to_pytorch': [], 'tokenization_funnel': ['FunnelTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Any = ['FunnelTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Optional[int] = [ 'FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST', 'FunnelBaseModel', 'FunnelForMaskedLM', 'FunnelForMultipleChoice', 'FunnelForPreTraining', 'FunnelForQuestionAnswering', 'FunnelForSequenceClassification', 'FunnelForTokenClassification', 'FunnelModel', 'FunnelPreTrainedModel', 'load_tf_weights_in_funnel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Tuple = [ 'TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFFunnelBaseModel', 'TFFunnelForMaskedLM', 'TFFunnelForMultipleChoice', 'TFFunnelForPreTraining', 'TFFunnelForQuestionAnswering', 'TFFunnelForSequenceClassification', 'TFFunnelForTokenClassification', 'TFFunnelModel', 'TFFunnelPreTrainedModel', ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys __lowerCAmelCase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
88
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer __A =logging.get_logger(__name__) __A ={'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} # See all BART models at https://huggingface.co/models?filter=bart __A ={ '''vocab_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/vocab.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/vocab.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json''', }, '''merges_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/merges.txt''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/merges.txt''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json''', }, } __A ={ '''facebook/bart-base''': 1_0_2_4, '''facebook/bart-large''': 1_0_2_4, '''facebook/bart-large-mnli''': 1_0_2_4, '''facebook/bart-large-cnn''': 1_0_2_4, '''facebook/bart-large-xsum''': 1_0_2_4, '''yjernite/bart_eli5''': 1_0_2_4, } class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ['input_ids', 'attention_mask'] lowerCAmelCase__ = BartTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase="replace" , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=False , lowercase=True , **lowercase , ) -> str: super().__init__( lowercase , lowercase , tokenizer_file=lowercase , errors=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , add_prefix_space=lowercase , trim_offsets=lowercase , **lowercase , ) lowerCamelCase_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowercase ) != add_prefix_space: lowerCamelCase_ = getattr(lowercase , pre_tok_state.pop("type" ) ) lowerCamelCase_ = add_prefix_space lowerCamelCase_ = pre_tok_class(**lowercase ) lowerCamelCase_ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCamelCase_ = "post_processor" lowerCamelCase_ = getattr(self.backend_tokenizer , lowercase , lowercase ) if tokenizer_component_instance: lowerCamelCase_ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase_ = tuple(state["sep"] ) if "cls" in state: lowerCamelCase_ = tuple(state["cls"] ) lowerCamelCase_ = False if state.get("add_prefix_space" , lowercase ) != add_prefix_space: lowerCamelCase_ = add_prefix_space lowerCamelCase_ = True if state.get("trim_offsets" , lowercase ) != trim_offsets: lowerCamelCase_ = trim_offsets lowerCamelCase_ = True if changes_to_apply: lowerCamelCase_ = getattr(lowercase , state.pop("type" ) ) lowerCamelCase_ = component_class(**lowercase ) setattr(self.backend_tokenizer , lowercase , lowercase ) @property def SCREAMING_SNAKE_CASE_( self ) -> str: if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Optional[Any]: lowerCamelCase_ = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase ) if isinstance(lowercase , lowercase ) else value lowerCamelCase_ = value def SCREAMING_SNAKE_CASE_( self , *lowercase , **lowercase ) -> BatchEncoding: lowerCamelCase_ = kwargs.get("is_split_into_words" , lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowercase , **lowercase ) def SCREAMING_SNAKE_CASE_( self , *lowercase , **lowercase ) -> BatchEncoding: lowerCamelCase_ = kwargs.get("is_split_into_words" , lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowercase , **lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> Tuple[str]: lowerCamelCase_ = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase=None ) -> Dict: lowerCamelCase_ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
47
import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter __A =True except ImportError: __A =False __A =logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCamelCase_ ( lowerCamelCase__ ): return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class _SCREAMING_SNAKE_CASE ( snake_case_ ): @staticmethod def SCREAMING_SNAKE_CASE_( lowercase ) -> int: lowerCamelCase_ = parser.add_parser("add-new-model" ) add_new_model_parser.add_argument("--testing" , action="store_true" , help="If in testing mode." ) add_new_model_parser.add_argument("--testing_file" , type=lowercase , help="Configuration file on which to run." ) add_new_model_parser.add_argument( "--path" , type=lowercase , help="Path to cookiecutter. Should only be used for testing purposes." ) add_new_model_parser.set_defaults(func=lowercase ) def __init__( self , lowercase , lowercase , lowercase=None , *lowercase ) -> List[str]: lowerCamelCase_ = testing lowerCamelCase_ = testing_file lowerCamelCase_ = path def SCREAMING_SNAKE_CASE_( self ) -> str: warnings.warn( "The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. " "It is not actively maintained anymore, so might give a result that won't pass all tests and quality " "checks, you should use `transformers-cli add-new-model-like` instead." ) if not _has_cookiecutter: raise ImportError( "Model creation dependencies are required to use the `add_new_model` command. Install them by running " "the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n" ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory lowerCamelCase_ = [directory for directory in os.listdir() if "cookiecutter-template-" == directory[:22]] if len(lowercase ) > 0: raise ValueError( "Several directories starting with `cookiecutter-template-` in current working directory. " "Please clean your directory by removing all folders starting with `cookiecutter-template-` or " "change your working directory." ) lowerCamelCase_ = ( Path(lowercase ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) lowerCamelCase_ = path_to_transformer_root / "templates" / "adding_a_new_model" # Execute cookiecutter if not self._testing: cookiecutter(str(lowercase ) ) else: with open(self._testing_file , "r" ) as configuration_file: lowerCamelCase_ = json.load(lowercase ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=lowercase , extra_context=lowercase , ) lowerCamelCase_ = [directory for directory in os.listdir() if "cookiecutter-template-" in directory[:22]][0] # Retrieve configuration with open(directory + "/configuration.json" , "r" ) as configuration_file: lowerCamelCase_ = json.load(lowercase ) lowerCamelCase_ = configuration["lowercase_modelname"] lowerCamelCase_ = configuration["generate_tensorflow_pytorch_and_flax"] os.remove(f'{directory}/configuration.json' ) lowerCamelCase_ = "PyTorch" in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = "TensorFlow" in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = "Flax" in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = f'{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}' os.makedirs(lowercase , exist_ok=lowercase ) os.makedirs(f'{path_to_transformer_root}/tests/models/{lowercase_model_name}' , exist_ok=lowercase ) # Tests require submodules as they have parent imports with open(f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py' , "w" ): pass shutil.move( f'{directory}/__init__.py' , f'{model_dir}/__init__.py' , ) shutil.move( f'{directory}/configuration_{lowercase_model_name}.py' , f'{model_dir}/configuration_{lowercase_model_name}.py' , ) def remove_copy_lines(lowercase ): with open(lowercase , "r" ) as f: lowerCamelCase_ = f.readlines() with open(lowercase , "w" ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(lowercase ) if output_pytorch: if not self._testing: remove_copy_lines(f'{directory}/modeling_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_{lowercase_model_name}.py' , f'{model_dir}/modeling_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_{lowercase_model_name}.py' ) if output_tensorflow: if not self._testing: remove_copy_lines(f'{directory}/modeling_tf_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_tf_{lowercase_model_name}.py' , f'{model_dir}/modeling_tf_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_tf_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_tf_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_tf_{lowercase_model_name}.py' ) if output_flax: if not self._testing: remove_copy_lines(f'{directory}/modeling_flax_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_flax_{lowercase_model_name}.py' , f'{model_dir}/modeling_flax_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_flax_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_flax_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_flax_{lowercase_model_name}.py' ) shutil.move( f'{directory}/{lowercase_model_name}.md' , f'{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md' , ) shutil.move( f'{directory}/tokenization_{lowercase_model_name}.py' , f'{model_dir}/tokenization_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/tokenization_fast_{lowercase_model_name}.py' , f'{model_dir}/tokenization_{lowercase_model_name}_fast.py' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(lowercase , lowercase , lowercase ): # Create temp file lowerCamelCase_ , lowerCamelCase_ = mkstemp() lowerCamelCase_ = False with fdopen(lowercase , "w" ) as new_file: with open(lowercase ) as old_file: for line in old_file: new_file.write(lowercase ) if line_to_copy_below in line: lowerCamelCase_ = True for line_to_copy in lines_to_copy: new_file.write(lowercase ) if not line_found: raise ValueError(f'Line {line_to_copy_below} was not found in file.' ) # Copy the file permissions from the old file to the new file copymode(lowercase , lowercase ) # Remove original file remove(lowercase ) # Move new file move(lowercase , lowercase ) def skip_units(lowercase ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(lowercase ): with open(lowercase ) as datafile: lowerCamelCase_ = [] lowerCamelCase_ = False lowerCamelCase_ = False for line in datafile: if "# To replace in: " in line and "##" not in line: lowerCamelCase_ = line.split("\"" )[1] lowerCamelCase_ = skip_units(lowercase ) elif "# Below: " in line and "##" not in line: lowerCamelCase_ = line.split("\"" )[1] lowerCamelCase_ = skip_units(lowercase ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(lowercase , lowercase , lowercase ) lowerCamelCase_ = [] elif "# Replace with" in line and "##" not in line: lowerCamelCase_ = [] elif "##" not in line: lines_to_copy.append(lowercase ) remove(lowercase ) replace_in_files(f'{directory}/to_replace_{lowercase_model_name}.py' ) os.rmdir(lowercase )
47
1
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version a : Optional[Any] = logging.getLogger(__name__) require_version('pytorch_lightning>=1.0.4') a : List[str] = { 'base': AutoModel, 'sequence-classification': AutoModelForSequenceClassification, 'question-answering': AutoModelForQuestionAnswering, 'pretraining': AutoModelForPreTraining, 'token-classification': AutoModelForTokenClassification, 'language-modeling': AutoModelWithLMHead, 'summarization': AutoModelForSeqaSeqLM, 'translation': AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization a : Any = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } a : str = sorted(arg_to_scheduler.keys()) a : Any = '{' + ', '.join(arg_to_scheduler_choices) + '}' class _a ( pl.LightningModule ): def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_="base", SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, **SCREAMING_SNAKE_CASE_, ) -> Union[str, Any]: super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[Any] = 0 UpperCAmelCase_: Any = Path(self.hparams.output_dir ) UpperCAmelCase_: Dict = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: UpperCAmelCase_: str = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path, **({"""num_labels""": num_labels} if num_labels is not None else {}), cache_dir=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_, ) else: UpperCAmelCase_: PretrainedConfig = config UpperCAmelCase_: Union[str, Any] = ("""encoder_layerdrop""", """decoder_layerdrop""", """dropout""", """attention_dropout""") for p in extra_model_params: if getattr(self.hparams, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): assert hasattr(self.config, SCREAMING_SNAKE_CASE_ ), f'model config doesn\'t have a `{p}` attribute' setattr(self.config, SCREAMING_SNAKE_CASE_, getattr(self.hparams, SCREAMING_SNAKE_CASE_ ) ) if tokenizer is None: UpperCAmelCase_: List[Any] = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path, cache_dir=SCREAMING_SNAKE_CASE_, ) else: UpperCAmelCase_: PreTrainedTokenizer = tokenizer UpperCAmelCase_: List[Any] = MODEL_MODES[mode] if model is None: UpperCAmelCase_: Any = self.model_type.from_pretrained( self.hparams.model_name_or_path, from_tf=bool(""".ckpt""" in self.hparams.model_name_or_path ), config=self.config, cache_dir=SCREAMING_SNAKE_CASE_, ) else: UpperCAmelCase_: Optional[Any] = model def __snake_case (self, *SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCAmelCase_: Any = self.model_type.from_pretrained(*SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> int: UpperCAmelCase_: Dict = arg_to_scheduler[self.hparams.lr_scheduler] UpperCAmelCase_: Optional[Any] = get_schedule_func( self.opt, num_warmup_steps=self.hparams.warmup_steps, num_training_steps=self.total_steps() ) UpperCAmelCase_: Dict = {"""scheduler""": scheduler, """interval""": """step""", """frequency""": 1} return scheduler def __snake_case (self ) -> Union[str, Any]: UpperCAmelCase_: str = self.model UpperCAmelCase_: str = ["""bias""", """LayerNorm.weight"""] UpperCAmelCase_: str = [ { """params""": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters """weight_decay""": self.hparams.weight_decay, }, { """params""": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] if self.hparams.adafactor: UpperCAmelCase_: List[str] = Adafactor( SCREAMING_SNAKE_CASE_, lr=self.hparams.learning_rate, scale_parameter=SCREAMING_SNAKE_CASE_, relative_step=SCREAMING_SNAKE_CASE_ ) else: UpperCAmelCase_: Union[str, Any] = AdamW( SCREAMING_SNAKE_CASE_, lr=self.hparams.learning_rate, eps=self.hparams.adam_epsilon ) UpperCAmelCase_: Optional[int] = optimizer UpperCAmelCase_: int = self.get_lr_scheduler() return [optimizer], [scheduler] def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: return self.validation_step(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Dict: return self.validation_end(SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> int: UpperCAmelCase_: Tuple = max(1, self.hparams.gpus ) # TODO: consider num_tpu_cores UpperCAmelCase_: int = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Optional[int]: if stage == "test": UpperCAmelCase_: int = len(self.test_dataloader().dataset ) else: UpperCAmelCase_: Dict = self.get_dataloader("""train""", self.hparams.train_batch_size, shuffle=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = len(self.train_dataloader().dataset ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = False ) -> str: raise NotImplementedError("""You must implement this for your task""" ) def __snake_case (self ) -> List[str]: return self.train_loader def __snake_case (self ) -> int: return self.get_dataloader("""dev""", self.hparams.eval_batch_size, shuffle=SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Any: return self.get_dataloader("""test""", self.hparams.eval_batch_size, shuffle=SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Dict: return os.path.join( self.hparams.data_dir, """cached_{}_{}_{}""".format( SCREAMING_SNAKE_CASE_, list(filter(SCREAMING_SNAKE_CASE_, self.hparams.model_name_or_path.split("""/""" ) ) ).pop(), str(self.hparams.max_seq_length ), ), ) @pl.utilities.rank_zero_only def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCAmelCase_: List[str] = self.output_dir.joinpath("""best_tfmr""" ) UpperCAmelCase_: List[Any] = self.step_count self.model.save_pretrained(SCREAMING_SNAKE_CASE_ ) self.tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) @staticmethod def __snake_case (SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[Any]: parser.add_argument( """--model_name_or_path""", default=SCREAMING_SNAKE_CASE_, type=SCREAMING_SNAKE_CASE_, required=SCREAMING_SNAKE_CASE_, help="""Path to pretrained model or model identifier from huggingface.co/models""", ) parser.add_argument( """--config_name""", default="""""", type=SCREAMING_SNAKE_CASE_, help="""Pretrained config name or path if not the same as model_name""" ) parser.add_argument( """--tokenizer_name""", default=SCREAMING_SNAKE_CASE_, type=SCREAMING_SNAKE_CASE_, help="""Pretrained tokenizer name or path if not the same as model_name""", ) parser.add_argument( """--cache_dir""", default=str(Path(SCREAMING_SNAKE_CASE_ ).parent / """test_run""" / """cache""" ), type=SCREAMING_SNAKE_CASE_, help="""Where do you want to store the pre-trained models downloaded from huggingface.co""", ) parser.add_argument( """--encoder_layerdrop""", type=SCREAMING_SNAKE_CASE_, help="""Encoder layer dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--decoder_layerdrop""", type=SCREAMING_SNAKE_CASE_, help="""Decoder layer dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--dropout""", type=SCREAMING_SNAKE_CASE_, help="""Dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--attention_dropout""", type=SCREAMING_SNAKE_CASE_, help="""Attention dropout probability (Optional). Goes into model.config""", ) parser.add_argument("""--learning_rate""", default=5E-5, type=SCREAMING_SNAKE_CASE_, help="""The initial learning rate for Adam.""" ) parser.add_argument( """--lr_scheduler""", default="""linear""", choices=SCREAMING_SNAKE_CASE_, metavar=SCREAMING_SNAKE_CASE_, type=SCREAMING_SNAKE_CASE_, help="""Learning rate scheduler""", ) parser.add_argument("""--weight_decay""", default=0.0, type=SCREAMING_SNAKE_CASE_, help="""Weight decay if we apply some.""" ) parser.add_argument("""--adam_epsilon""", default=1E-8, type=SCREAMING_SNAKE_CASE_, help="""Epsilon for Adam optimizer.""" ) parser.add_argument("""--warmup_steps""", default=0, type=SCREAMING_SNAKE_CASE_, help="""Linear warmup over warmup_steps.""" ) parser.add_argument("""--num_workers""", default=4, type=SCREAMING_SNAKE_CASE_, help="""kwarg passed to DataLoader""" ) parser.add_argument("""--num_train_epochs""", dest="""max_epochs""", default=3, type=SCREAMING_SNAKE_CASE_ ) parser.add_argument("""--train_batch_size""", default=32, type=SCREAMING_SNAKE_CASE_ ) parser.add_argument("""--eval_batch_size""", default=32, type=SCREAMING_SNAKE_CASE_ ) parser.add_argument("""--adafactor""", action="""store_true""" ) class _a ( pl.Callback ): def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[str]: if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class _a ( pl.Callback ): def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(SCREAMING_SNAKE_CASE_ ) class _a ( pl.Callback ): def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> int: UpperCAmelCase_: Optional[Any] = trainer.lr_schedulers[0]["""scheduler"""] UpperCAmelCase_: Optional[int] = {f'lr_group_{i}': lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: rank_zero_info("""***** Validation results *****""" ) UpperCAmelCase_: int = trainer.callback_metrics # Log results for key in sorted(SCREAMING_SNAKE_CASE_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(SCREAMING_SNAKE_CASE_, str(metrics[key] ) ) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: rank_zero_info("""***** Test results *****""" ) UpperCAmelCase_: Any = trainer.callback_metrics # Log and save results to file UpperCAmelCase_: List[Any] = os.path.join(pl_module.hparams.output_dir, """test_results.txt""" ) with open(SCREAMING_SNAKE_CASE_, """w""" ) as writer: for key in sorted(SCREAMING_SNAKE_CASE_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(SCREAMING_SNAKE_CASE_, str(metrics[key] ) ) ) writer.write("""{} = {}\n""".format(SCREAMING_SNAKE_CASE_, str(metrics[key] ) ) ) def lowerCAmelCase_ (lowerCAmelCase__: int , lowerCAmelCase__: Optional[int] ): """simple docstring""" parser.add_argument( """--output_dir""" , default=str(Path(lowerCAmelCase__ ).parent / """test_run""" / """model_checkpoints""" ) , type=lowerCAmelCase__ , help="""The output directory where the model predictions and checkpoints will be written.""" , ) parser.add_argument( """--fp16""" , action="""store_true""" , help="""Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit""" , ) parser.add_argument( """--fp16_opt_level""" , type=lowerCAmelCase__ , default="""O2""" , help=( """For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3'].""" """See details at https://nvidia.github.io/apex/amp.html""" ) , ) parser.add_argument("""--n_tpu_cores""" , dest="""tpu_cores""" , type=lowerCAmelCase__ ) parser.add_argument("""--max_grad_norm""" , dest="""gradient_clip_val""" , default=1.0 , type=lowerCAmelCase__ , help="""Max gradient norm""" ) parser.add_argument("""--do_train""" , action="""store_true""" , help="""Whether to run training.""" ) parser.add_argument("""--do_predict""" , action="""store_true""" , help="""Whether to run predictions on the test set.""" ) parser.add_argument( """--gradient_accumulation_steps""" , dest="""accumulate_grad_batches""" , type=lowerCAmelCase__ , default=1 , help="""Number of updates steps to accumulate before performing a backward/update pass.""" , ) parser.add_argument("""--seed""" , type=lowerCAmelCase__ , default=4_2 , help="""random seed for initialization""" ) parser.add_argument( """--data_dir""" , default=str(Path(lowerCAmelCase__ ).parent / """test_run""" / """dummy-train-data""" ) , type=lowerCAmelCase__ , help="""The input data dir. Should contain the training files for the CoNLL-2003 NER task.""" , ) def lowerCAmelCase_ (lowerCAmelCase__: BaseTransformer , lowerCAmelCase__: argparse.Namespace , lowerCAmelCase__: Union[str, Any]=None , lowerCAmelCase__: Optional[Any]=True , lowerCAmelCase__: Dict=[] , lowerCAmelCase__: Tuple=None , lowerCAmelCase__: List[str]=None , **lowerCAmelCase__: List[Any] , ): """simple docstring""" pl.seed_everything(args.seed ) # init model UpperCAmelCase_: Dict = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=lowerCAmelCase__ ) # add custom checkpoints if checkpoint_callback is None: UpperCAmelCase_: Dict = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="""checkpoint""" , monitor="""val_loss""" , mode="""min""" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(lowerCAmelCase__ ) if logging_callback is None: UpperCAmelCase_: Any = LoggingCallback() UpperCAmelCase_: Optional[int] = {} if args.fpaa: UpperCAmelCase_: List[str] = 1_6 if args.gpus > 1: UpperCAmelCase_: str = """auto""" UpperCAmelCase_: Union[str, Any] = """ddp""" UpperCAmelCase_: Tuple = args.accumulate_grad_batches UpperCAmelCase_: Optional[int] = None UpperCAmelCase_: List[Any] = """auto""" UpperCAmelCase_: Any = pl.Trainer.from_argparse_args( lowerCAmelCase__ , weights_summary=lowerCAmelCase__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=lowerCAmelCase__ , val_check_interval=1 , num_sanity_val_steps=2 , **lowerCAmelCase__ , ) if args.do_train: trainer.fit(lowerCAmelCase__ ) else: print("""RAG modeling tests with new set functions successfuly executed!""" ) return trainer
147
from torch import nn def lowerCAmelCase_ (lowerCAmelCase__: Optional[int] ): """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'Unsupported activation function: {act_fn}' )
147
1
"""simple docstring""" # This is the module that test_patching.py uses to test patch_submodule() import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests _lowerCAmelCase : int = open # noqa: we just need to have a builtin inside this module to test it properly
202
"""simple docstring""" from __future__ import annotations from typing import Any def __snake_case ( SCREAMING_SNAKE_CASE__ : list[Any] ) -> None: '''simple docstring''' create_state_space_tree(SCREAMING_SNAKE_CASE__ , [] , 0 ) def __snake_case ( SCREAMING_SNAKE_CASE__ : list[Any] , SCREAMING_SNAKE_CASE__ : list[Any] , SCREAMING_SNAKE_CASE__ : int ) -> None: '''simple docstring''' if index == len(SCREAMING_SNAKE_CASE__ ): print(SCREAMING_SNAKE_CASE__ ) return create_state_space_tree(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _lowerCAmelCase : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
202
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=3 , lowercase=30 , lowercase=400 , lowercase=True , lowercase=None , lowercase=True , lowercase=[0.5, 0.5, 0.5] , lowercase=[0.5, 0.5, 0.5] , lowercase=True , lowercase=1 / 255 , lowercase=True , ) -> Optional[int]: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowerCamelCase_ = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_pad def SCREAMING_SNAKE_CASE_( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase=False ) -> Optional[Any]: if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(lowercase , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size["shortest_edge"] * h / w ) lowerCamelCase_ = self.size["shortest_edge"] elif w > h: lowerCamelCase_ = self.size["shortest_edge"] lowerCamelCase_ = int(self.size["shortest_edge"] * w / h ) else: lowerCamelCase_ = self.size["shortest_edge"] lowerCamelCase_ = self.size["shortest_edge"] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(lowercase , key=lambda lowercase : item[0] )[0] lowerCamelCase_ = max(lowercase , key=lambda lowercase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( snake_case_ , unittest.TestCase ): lowerCAmelCase__ = YolosImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = YolosImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE_( self ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , "image_mean" ) ) self.assertTrue(hasattr(lowercase , "image_std" ) ) self.assertTrue(hasattr(lowercase , "do_normalize" ) ) self.assertTrue(hasattr(lowercase , "do_resize" ) ) self.assertTrue(hasattr(lowercase , "size" ) ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowercase ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowercase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: pass def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) lowerCamelCase_ = image_processing(lowercase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE_( self ) -> Any: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(lowercase , return_tensors="pt" ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(lowercase , return_tensors="pt" ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: # Initialize image_processings lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) lowerCamelCase_ = self.image_processing_class(do_resize=lowercase , do_normalize=lowercase , do_rescale=lowercase ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowerCamelCase_ = image_processing_a.pad(lowercase , return_tensors="pt" ) lowerCamelCase_ = image_processing_a(lowercase , return_tensors="pt" ) self.assertTrue( torch.allclose(encoded_images_with_method["pixel_values"] , encoded_images["pixel_values"] , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE_( self ) -> int: # prepare image and target lowerCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {"image_id": 39769, "annotations": target} # encode them lowerCamelCase_ = YolosImageProcessor.from_pretrained("hustvl/yolos-small" ) lowerCamelCase_ = image_processing(images=lowercase , annotations=lowercase , return_tensors="pt" ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowercase ) lowerCamelCase_ = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowercase , atol=1e-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowercase ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowercase ) lowerCamelCase_ = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowercase , atol=1e-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowercase ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowercase ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowercase ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowercase ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowercase ) ) @slow def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: # prepare image, target and masks_path lowerCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} lowerCamelCase_ = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them lowerCamelCase_ = YolosImageProcessor(format="coco_panoptic" ) lowerCamelCase_ = image_processing(images=lowercase , annotations=lowercase , masks_path=lowercase , return_tensors="pt" ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowercase ) lowerCamelCase_ = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowercase , atol=1e-4 ) ) # verify area lowerCamelCase_ = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowercase ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowercase ) lowerCamelCase_ = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowercase , atol=1e-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowercase ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowercase ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowercase ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowercase ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowercase ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowercase ) )
19
import warnings warnings.warn( 'memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ' '`from accelerate import find_executable_batch_size` to avoid this warning.', FutureWarning, )
207
0
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa SCREAMING_SNAKE_CASE : Union[str, Any] = logging.getLogger(__name__) class UpperCamelCase ( __lowerCamelCase ): '''simple docstring''' lowercase : Any ='summarization' lowercase : List[str] =['loss'] lowercase : Union[str, Any] =ROUGE_KEYS lowercase : List[str] ='rouge2' def __init__( self , UpperCamelCase_ , **UpperCamelCase_ ): if hparams.sortish_sampler and hparams.gpus > 1: lowercase_ :str = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError('''Dynamic Batch size does not work for multi-gpu training''' ) if hparams.sortish_sampler: raise ValueError('''--sortish_sampler and --max_tokens_per_batch may not be used simultaneously''' ) super().__init__(UpperCamelCase_ , num_labels=UpperCamelCase_ , mode=self.mode , **UpperCamelCase_ ) use_task_specific_params(self.model , '''summarization''' ) save_git_info(self.hparams.output_dir ) lowercase_ :Optional[int] = Path(self.output_dir ) / '''metrics.json''' lowercase_ :Dict = Path(self.output_dir ) / '''hparams.pkl''' pickle_save(self.hparams , self.hparams_save_path ) lowercase_ :Tuple = 0 lowercase_ :Tuple = defaultdict(UpperCamelCase_ ) lowercase_ :Any = self.config.model_type lowercase_ :Optional[int] = self.config.tgt_vocab_size if self.model_type == '''fsmt''' else self.config.vocab_size lowercase_ :List[Any] = { '''data_dir''': self.hparams.data_dir, '''max_source_length''': self.hparams.max_source_length, '''prefix''': self.model.config.prefix or '''''', } lowercase_ :Optional[int] = { '''train''': self.hparams.n_train, '''val''': self.hparams.n_val, '''test''': self.hparams.n_test, } lowercase_ :Tuple = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} lowercase_ :List[Any] = { '''train''': self.hparams.max_target_length, '''val''': self.hparams.val_max_target_length, '''test''': self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], f"target_lens: {self.target_lens}" assert self.target_lens["train"] <= self.target_lens["test"], f"target_lens: {self.target_lens}" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) lowercase_ :List[str] = get_git_info()['''repo_sha'''] lowercase_ :List[Any] = hparams.num_workers lowercase_ :int = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , UpperCamelCase_ ): lowercase_ :Any = self.tokenizer.lang_code_to_id[hparams.tgt_lang] lowercase_ :List[str] = self.decoder_start_token_id lowercase_ :List[str] = ( SeqaSeqDataset if hasattr(self.tokenizer , '''prepare_seq2seq_batch''' ) else LegacySeqaSeqDataset ) lowercase_ :Optional[Any] = False lowercase_ :List[Any] = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: lowercase_ :Dict = self.hparams.eval_max_gen_length else: lowercase_ :Union[str, Any] = self.model.config.max_length lowercase_ :List[Any] = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Dict = { k: self.tokenizer.batch_decode(v.tolist() ) if '''mask''' not in k else v.shape for k, v in batch.items() } save_json(UpperCamelCase_ , Path(self.output_dir ) / '''text_batch.json''' ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / '''tok_batch.json''' ) lowercase_ :Union[str, Any] = True return readable_batch def UpperCamelCase ( self , UpperCamelCase_ , **UpperCamelCase_ ): return self.model(UpperCamelCase_ , **UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Optional[int] = self.tokenizer.batch_decode( UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ ) return lmap(str.strip , UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Any = self.tokenizer.pad_token_id lowercase_ , lowercase_ :int = batch['''input_ids'''], batch['''attention_mask'''] lowercase_ :int = batch['''labels'''] if isinstance(self.model , UpperCamelCase_ ): lowercase_ :Optional[int] = self.model._shift_right(UpperCamelCase_ ) else: lowercase_ :Any = shift_tokens_right(UpperCamelCase_ , UpperCamelCase_ ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero lowercase_ :int = decoder_input_ids self.save_readable_batch(UpperCamelCase_ ) lowercase_ :Any = self(UpperCamelCase_ , attention_mask=UpperCamelCase_ , decoder_input_ids=UpperCamelCase_ , use_cache=UpperCamelCase_ ) lowercase_ :Union[str, Any] = outputs['''logits'''] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id lowercase_ :str = nn.CrossEntropyLoss(ignore_index=UpperCamelCase_ ) assert lm_logits.shape[-1] == self.vocab_size lowercase_ :int = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: lowercase_ :Dict = nn.functional.log_softmax(UpperCamelCase_ , dim=-1 ) lowercase_ , lowercase_ :Optional[int] = label_smoothed_nll_loss( UpperCamelCase_ , UpperCamelCase_ , self.hparams.label_smoothing , ignore_index=UpperCamelCase_ ) return (loss,) @property def UpperCamelCase ( self ): return self.tokenizer.pad_token_id def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :Dict = self._step(UpperCamelCase_ ) lowercase_ :List[str] = dict(zip(self.loss_names , UpperCamelCase_ ) ) # tokens per batch lowercase_ :Union[str, Any] = batch['''input_ids'''].ne(self.pad ).sum() + batch['''labels'''].ne(self.pad ).sum() lowercase_ :Tuple = batch['''input_ids'''].shape[0] lowercase_ :Dict = batch['''input_ids'''].eq(self.pad ).sum() lowercase_ :List[Any] = batch['''input_ids'''].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): return self._generative_step(UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_="val" ): self.step_count += 1 lowercase_ :Any = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} lowercase_ :int = losses['''loss'''] lowercase_ :Dict = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ['''gen_time''', '''gen_len'''] } lowercase_ :Optional[Any] = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) lowercase_ :List[str] = torch.tensor(UpperCamelCase_ ).type_as(UpperCamelCase_ ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(UpperCamelCase_ ) lowercase_ :Dict = {f"{prefix}_avg_{k}": x for k, x in losses.items()} lowercase_ :str = self.step_count self.metrics[prefix].append(UpperCamelCase_ ) # callback writes this to self.metrics_save_path lowercase_ :List[str] = flatten_list([x['''preds'''] for x in outputs] ) return { "log": all_metrics, "preds": preds, f"{prefix}_loss": loss, f"{prefix}_{self.val_metric}": metric_tensor, } def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): return calculate_rouge(UpperCamelCase_ , UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :str = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') lowercase_ :Optional[Any] = self.model.generate( batch['''input_ids'''] , attention_mask=batch['''attention_mask'''] , use_cache=UpperCamelCase_ , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) lowercase_ :str = (time.time() - ta) / batch['''input_ids'''].shape[0] lowercase_ :int = self.ids_to_clean_text(UpperCamelCase_ ) lowercase_ :Any = self.ids_to_clean_text(batch['''labels'''] ) lowercase_ :List[Any] = self._step(UpperCamelCase_ ) lowercase_ :Dict = dict(zip(self.loss_names , UpperCamelCase_ ) ) lowercase_ :Tuple = self.calc_generative_metrics(UpperCamelCase_ , UpperCamelCase_ ) lowercase_ :Optional[Any] = np.mean(lmap(UpperCamelCase_ , UpperCamelCase_ ) ) base_metrics.update(gen_time=UpperCamelCase_ , gen_len=UpperCamelCase_ , preds=UpperCamelCase_ , target=UpperCamelCase_ , **UpperCamelCase_ ) return base_metrics def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): return self._generative_step(UpperCamelCase_ ) def UpperCamelCase ( self , UpperCamelCase_ ): return self.validation_epoch_end(UpperCamelCase_ , prefix='''test''' ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :Dict = self.n_obs[type_path] lowercase_ :int = self.target_lens[type_path] lowercase_ :Dict = self.dataset_class( self.tokenizer , type_path=UpperCamelCase_ , n_obs=UpperCamelCase_ , max_target_length=UpperCamelCase_ , **self.dataset_kwargs , ) return dataset def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = False ): lowercase_ :List[Any] = self.get_dataset(UpperCamelCase_ ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": lowercase_ :Optional[Any] = dataset.make_sortish_sampler(UpperCamelCase_ , distributed=self.hparams.gpus > 1 ) return DataLoader( UpperCamelCase_ , batch_size=UpperCamelCase_ , collate_fn=dataset.collate_fn , shuffle=UpperCamelCase_ , num_workers=self.num_workers , sampler=UpperCamelCase_ , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": lowercase_ :Any = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( UpperCamelCase_ , batch_sampler=UpperCamelCase_ , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( UpperCamelCase_ , batch_size=UpperCamelCase_ , collate_fn=dataset.collate_fn , shuffle=UpperCamelCase_ , num_workers=self.num_workers , sampler=UpperCamelCase_ , ) def UpperCamelCase ( self ): lowercase_ :Optional[int] = self.get_dataloader('''train''' , batch_size=self.hparams.train_batch_size , shuffle=UpperCamelCase_ ) return dataloader def UpperCamelCase ( self ): return self.get_dataloader('''val''' , batch_size=self.hparams.eval_batch_size ) def UpperCamelCase ( self ): return self.get_dataloader('''test''' , batch_size=self.hparams.eval_batch_size ) @staticmethod def UpperCamelCase ( UpperCamelCase_ , UpperCamelCase_ ): BaseTransformer.add_model_specific_args(UpperCamelCase_ , UpperCamelCase_ ) add_generic_args(UpperCamelCase_ , UpperCamelCase_ ) parser.add_argument( '''--max_source_length''' , default=1024 , type=UpperCamelCase_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--max_target_length''' , default=56 , type=UpperCamelCase_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--val_max_target_length''' , default=142 , type=UpperCamelCase_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--test_max_target_length''' , default=142 , type=UpperCamelCase_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument('''--freeze_encoder''' , action='''store_true''' ) parser.add_argument('''--freeze_embeds''' , action='''store_true''' ) parser.add_argument('''--sortish_sampler''' , action='''store_true''' , default=UpperCamelCase_ ) parser.add_argument('''--overwrite_output_dir''' , action='''store_true''' , default=UpperCamelCase_ ) parser.add_argument('''--max_tokens_per_batch''' , type=UpperCamelCase_ , default=UpperCamelCase_ ) parser.add_argument('''--logger_name''' , type=UpperCamelCase_ , choices=['''default''', '''wandb''', '''wandb_shared'''] , default='''default''' ) parser.add_argument('''--n_train''' , type=UpperCamelCase_ , default=-1 , required=UpperCamelCase_ , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_val''' , type=UpperCamelCase_ , default=500 , required=UpperCamelCase_ , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_test''' , type=UpperCamelCase_ , default=-1 , required=UpperCamelCase_ , help='''# examples. -1 means use all.''' ) parser.add_argument( '''--task''' , type=UpperCamelCase_ , default='''summarization''' , required=UpperCamelCase_ , help='''# examples. -1 means use all.''' ) parser.add_argument('''--label_smoothing''' , type=UpperCamelCase_ , default=0.0 , required=UpperCamelCase_ ) parser.add_argument('''--src_lang''' , type=UpperCamelCase_ , default='''''' , required=UpperCamelCase_ ) parser.add_argument('''--tgt_lang''' , type=UpperCamelCase_ , default='''''' , required=UpperCamelCase_ ) parser.add_argument('''--eval_beams''' , type=UpperCamelCase_ , default=UpperCamelCase_ , required=UpperCamelCase_ ) parser.add_argument( '''--val_metric''' , type=UpperCamelCase_ , default=UpperCamelCase_ , required=UpperCamelCase_ , choices=['''bleu''', '''rouge2''', '''loss''', None] ) parser.add_argument('''--eval_max_gen_length''' , type=UpperCamelCase_ , default=UpperCamelCase_ , help='''never generate more than n tokens''' ) parser.add_argument('''--save_top_k''' , type=UpperCamelCase_ , default=1 , required=UpperCamelCase_ , help='''How many checkpoints to save''' ) parser.add_argument( '''--early_stopping_patience''' , type=UpperCamelCase_ , default=-1 , required=UpperCamelCase_ , help=( '''-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So''' ''' val_check_interval will effect it.''' ) , ) return parser class UpperCamelCase ( __lowerCamelCase ): '''simple docstring''' lowercase : Optional[int] ='translation' lowercase : Any =['loss'] lowercase : List[str] =['bleu'] lowercase : List[str] ='bleu' def __init__( self , UpperCamelCase_ , **UpperCamelCase_ ): super().__init__(UpperCamelCase_ , **UpperCamelCase_ ) lowercase_ :Tuple = hparams.src_lang lowercase_ :Tuple = hparams.tgt_lang def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ): return calculate_bleu(UpperCamelCase_ , UpperCamelCase_ ) def UpperCamelCase ( _a , _a=None ) -> Tuple: '''simple docstring''' Path(args.output_dir ).mkdir(exist_ok=A__ ) check_output_dir(A__ , expected_items=3 ) if model is None: if "summarization" in args.task: lowercase_ :Dict = SummarizationModule(A__ ) else: lowercase_ :Optional[Any] = TranslationModule(A__ ) lowercase_ :List[Any] = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith('''/tmp''' ) or str(args.output_dir ).startswith('''/var''' ) ): lowercase_ :List[str] = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger lowercase_ :Dict = os.environ.get('''WANDB_PROJECT''' , A__ ) lowercase_ :Any = WandbLogger(name=model.output_dir.name , project=A__ ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger lowercase_ :Union[str, Any] = WandbLogger(name=model.output_dir.name , project=f"hf_{dataset}" ) if args.early_stopping_patience >= 0: lowercase_ :List[Any] = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: lowercase_ :Tuple = False lowercase_ :Tuple = args.val_metric == '''loss''' lowercase_ :str = generic_train( A__ , A__ , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , A__ ) , early_stopping_callback=A__ , logger=A__ , ) pickle_save(model.hparams , model.output_dir / '''hparams.pkl''' ) if not args.do_predict: return model lowercase_ :List[Any] = '''''' lowercase_ :Union[str, Any] = sorted(glob.glob(os.path.join(args.output_dir , '''*.ckpt''' ) , recursive=A__ ) ) if checkpoints: lowercase_ :int = checkpoints[-1] lowercase_ :int = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": SCREAMING_SNAKE_CASE : Dict = argparse.ArgumentParser() SCREAMING_SNAKE_CASE : Union[str, Any] = pl.Trainer.add_argparse_args(parser) SCREAMING_SNAKE_CASE : int = SummarizationModule.add_model_specific_args(parser, os.getcwd()) SCREAMING_SNAKE_CASE : Any = parser.parse_args() main(args)
357
import os import re import warnings from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_ta import TaTokenizer else: SCREAMING_SNAKE_CASE : Union[str, Any] = None SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Dict = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE : Any = { "vocab_file": { "t5-small": "https://huggingface.co/t5-small/resolve/main/spiece.model", "t5-base": "https://huggingface.co/t5-base/resolve/main/spiece.model", "t5-large": "https://huggingface.co/t5-large/resolve/main/spiece.model", "t5-3b": "https://huggingface.co/t5-3b/resolve/main/spiece.model", "t5-11b": "https://huggingface.co/t5-11b/resolve/main/spiece.model", }, "tokenizer_file": { "t5-small": "https://huggingface.co/t5-small/resolve/main/tokenizer.json", "t5-base": "https://huggingface.co/t5-base/resolve/main/tokenizer.json", "t5-large": "https://huggingface.co/t5-large/resolve/main/tokenizer.json", "t5-3b": "https://huggingface.co/t5-3b/resolve/main/tokenizer.json", "t5-11b": "https://huggingface.co/t5-11b/resolve/main/tokenizer.json", }, } # TODO(PVP) - this should be removed in Transformers v5 SCREAMING_SNAKE_CASE : Tuple = { "t5-small": 512, "t5-base": 512, "t5-large": 512, "t5-3b": 512, "t5-11b": 512, } class UpperCamelCase ( lowercase__ ): '''simple docstring''' lowercase : Tuple =VOCAB_FILES_NAMES lowercase : Dict =PRETRAINED_VOCAB_FILES_MAP lowercase : List[str] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase : Optional[Any] =["""input_ids""", """attention_mask"""] lowercase : str =TaTokenizer lowercase : List[int] =[] def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_="</s>" , UpperCamelCase_="<unk>" , UpperCamelCase_="<pad>" , UpperCamelCase_=100 , UpperCamelCase_=None , **UpperCamelCase_ , ): # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: lowercase_ :Union[str, Any] = [f"<extra_id_{i}>" for i in range(UpperCamelCase_ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra special tokens lowercase_ :Tuple = len(set(filter(lambda UpperCamelCase_ : bool('''extra_id_''' in str(UpperCamelCase_ ) ) , UpperCamelCase_ ) ) ) if extra_tokens != extra_ids: raise ValueError( f"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" ''' provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids''' ''' tokens''' ) super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) lowercase_ :Union[str, Any] = vocab_file lowercase_ :Optional[int] = False if not self.vocab_file else True lowercase_ :Dict = extra_ids @staticmethod def UpperCamelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): if pretrained_model_name_or_path in TaTokenizerFast.max_model_input_sizes: lowercase_ :Optional[int] = TaTokenizerFast.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( '''This tokenizer was incorrectly instantiated with a model max length of''' f" {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this" ''' behavior is kept to avoid breaking backwards compatibility when padding/encoding with''' ''' `truncation is True`.\n- Be aware that you SHOULD NOT rely on''' f" {pretrained_model_name_or_path} automatically truncating your input to" f" {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences" f" longer than {deprecated_max_model_length} you can either instantiate this tokenizer with" ''' `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please''' ''' instantiate this tokenizer with `model_max_length` set to your preferred value.''' , UpperCamelCase_ , ) return max_model_length def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(UpperCamelCase_ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return lowercase_ :Dict = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ): copyfile(self.vocab_file , UpperCamelCase_ ) logger.info(f"Copy vocab file to {out_vocab_file}" ) return (out_vocab_file,) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = None ): lowercase_ :Any = token_ids_a + [self.eos_token_id] if token_ids_a is None: return self.prefix_tokens + token_ids_a else: lowercase_ :str = token_ids_a + [self.eos_token_id] return self.prefix_tokens + token_ids_a + token_ids_a def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = None ): lowercase_ :Dict = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def UpperCamelCase ( self ): return list( set(filter(lambda UpperCamelCase_ : bool(re.search(R'''<extra_id_\d+>''' , UpperCamelCase_ ) ) is not None , self.additional_special_tokens ) ) ) def UpperCamelCase ( self ): return [self.convert_tokens_to_ids(UpperCamelCase_ ) for token in self.get_sentinel_tokens()]
252
0
"""simple docstring""" import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase=13 ,__UpperCamelCase=32 ,__UpperCamelCase=3 ,__UpperCamelCase=4 ,__UpperCamelCase=[10, 20, 30, 40] ,__UpperCamelCase=[2, 2, 3, 2] ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=37 ,__UpperCamelCase="gelu" ,__UpperCamelCase=10 ,__UpperCamelCase=0.02 ,__UpperCamelCase=["stage2", "stage3", "stage4"] ,__UpperCamelCase=[2, 3, 4] ,__UpperCamelCase=None ,) -> Any: '''simple docstring''' lowercase_ : Optional[int] = parent lowercase_ : Dict = batch_size lowercase_ : List[Any] = image_size lowercase_ : int = num_channels lowercase_ : List[Any] = num_stages lowercase_ : Optional[int] = hidden_sizes lowercase_ : str = depths lowercase_ : int = is_training lowercase_ : List[Any] = use_labels lowercase_ : Optional[int] = intermediate_size lowercase_ : str = hidden_act lowercase_ : Tuple = num_labels lowercase_ : Optional[int] = initializer_range lowercase_ : str = out_features lowercase_ : Any = out_indices lowercase_ : Optional[Any] = scope def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' lowercase_ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : Optional[Any] = ids_tensor([self.batch_size] ,self.num_labels ) lowercase_ : int = self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,num_stages=self.num_stages ,hidden_act=self.hidden_act ,is_decoder=__UpperCamelCase ,initializer_range=self.initializer_range ,out_features=self.out_features ,out_indices=self.out_indices ,num_labels=self.num_labels ,) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> str: '''simple docstring''' lowercase_ : Optional[int] = ConvNextModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Any = model(__UpperCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[Any]: '''simple docstring''' lowercase_ : str = ConvNextForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Any = model(__UpperCamelCase ,labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : int = ConvNextBackbone(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Tuple = model(__UpperCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) ,len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) ,[self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) ,len(config.out_features ) ) self.parent.assertListEqual(model.channels ,config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowercase_ : str = None lowercase_ : List[Any] = ConvNextBackbone(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : str = model(__UpperCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) ,1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) ,[self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) ,1 ) self.parent.assertListEqual(model.channels ,[config.hidden_sizes[-1]] ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : List[Any] = config_and_inputs lowercase_ : Any = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( lowercase_ , lowercase_ , unittest.TestCase ): lowercase = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) lowercase = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) lowercase = True lowercase = False lowercase = False lowercase = False lowercase = False def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : Optional[int] = ConvNextModelTester(self ) lowercase_ : Optional[int] = ConfigTester(self ,config_class=__UpperCamelCase ,has_text_modality=__UpperCamelCase ,hidden_size=37 ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _UpperCAmelCase ( self ) -> str: '''simple docstring''' return @unittest.skip(reason='ConvNext does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason='ConvNext does not support input and output embeddings' ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' pass @unittest.skip(reason='ConvNext does not use feedforward chunking' ) def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' pass def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ , lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : str = model_class(__UpperCamelCase ) lowercase_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : Optional[Any] = [*signature.parameters.keys()] lowercase_ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' def check_hidden_states_output(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ): lowercase_ : Dict = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): lowercase_ : List[str] = model(**self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) lowercase_ : Dict = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(__UpperCamelCase ) ,expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 4, self.model_tester.image_size // 4] ,) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : Any = True check_hidden_states_output(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : str = True check_hidden_states_output(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' lowercase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Optional[Any] = ConvNextModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def lowercase__( ): lowercase_ : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase ( unittest.TestCase ): @cached_property def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' return AutoImageProcessor.from_pretrained('facebook/convnext-tiny-224' ) if is_vision_available() else None @slow def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : List[str] = ConvNextForImageClassification.from_pretrained('facebook/convnext-tiny-224' ).to(__UpperCamelCase ) lowercase_ : Tuple = self.default_image_processor lowercase_ : List[Any] = prepare_img() lowercase_ : Union[str, Any] = image_processor(images=__UpperCamelCase ,return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): lowercase_ : Optional[Any] = model(**__UpperCamelCase ) # verify the logits lowercase_ : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,__UpperCamelCase ) lowercase_ : List[Any] = torch.tensor([-0.0260, -0.4739, 0.1911] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,__UpperCamelCase ,atol=1e-4 ) ) @require_torch class UpperCamelCase ( unittest.TestCase , lowercase_ ): lowercase = (ConvNextBackbone,) if is_torch_available() else () lowercase = ConvNextConfig lowercase = False def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : List[str] = ConvNextModelTester(self )
213
"""simple docstring""" from __future__ import annotations def lowercase__( __SCREAMING_SNAKE_CASE : list ): if not nums: raise ValueError('List is empty' ) return sum(__SCREAMING_SNAKE_CASE ) / len(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
213
1
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis_float32 (there's also the fix_lavis branch) # also note: to convert Vicuna checkpoints, we had to include /home/niels/python_projects/checkpoints/FastChat/vicuna-7b in lavis/configs/models/blip2/blip2_instruct_vicuna7b.yaml # same for Vicuna-13b from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipImageProcessor, InstructBlipConfig, InstructBlipForConditionalGeneration, InstructBlipProcessor, InstructBlipQFormerConfig, InstructBlipVisionConfig, LlamaConfig, LlamaTokenizerFast, TaConfig, TaTokenizerFast, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """https://raw.githubusercontent.com/salesforce/LAVIS/main/docs/_static/Confusing-Pictures.jpg""" _lowerCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ).convert("""RGB""" ) return image def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] # fmt: off # vision encoder rename_keys.append(("""visual_encoder.cls_token""", """vision_model.embeddings.class_embedding""") ) rename_keys.append(("""visual_encoder.pos_embed""", """vision_model.embeddings.position_embedding""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.weight""", """vision_model.embeddings.patch_embedding.weight""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.bias""", """vision_model.embeddings.patch_embedding.bias""") ) rename_keys.append(("""ln_vision.weight""", """vision_model.post_layernorm.weight""") ) rename_keys.append(("""ln_vision.bias""", """vision_model.post_layernorm.bias""") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((f"visual_encoder.blocks.{i}.norm1.weight", f"vision_model.encoder.layers.{i}.layer_norm1.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm1.bias", f"vision_model.encoder.layers.{i}.layer_norm1.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm2.weight", f"vision_model.encoder.layers.{i}.layer_norm2.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm2.bias", f"vision_model.encoder.layers.{i}.layer_norm2.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.qkv.weight", f"vision_model.encoder.layers.{i}.self_attn.qkv.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.proj.weight", f"vision_model.encoder.layers.{i}.self_attn.projection.weight",) ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.proj.bias", f"vision_model.encoder.layers.{i}.self_attn.projection.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc1.weight", f"vision_model.encoder.layers.{i}.mlp.fc1.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc1.bias", f"vision_model.encoder.layers.{i}.mlp.fc1.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc2.weight", f"vision_model.encoder.layers.{i}.mlp.fc2.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc2.bias", f"vision_model.encoder.layers.{i}.mlp.fc2.bias") ) # QFormer rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.weight""", """qformer.embeddings.layernorm.weight""") ) rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.bias""", """qformer.embeddings.layernorm.bias""") ) # fmt: on return rename_keys def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = dct.pop(lowerCAmelCase ) _lowerCAmelCase = val def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases _lowerCAmelCase = state_dict.pop(f"visual_encoder.blocks.{i}.attn.q_bias" ) _lowerCAmelCase = state_dict.pop(f"visual_encoder.blocks.{i}.attn.v_bias" ) # next, set bias in the state dict _lowerCAmelCase = torch.cat((q_bias, torch.zeros_like(lowerCAmelCase , requires_grad=lowerCAmelCase ), v_bias) ) _lowerCAmelCase = qkv_bias def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = 3_64 if """coco""" in model_name else 2_24 _lowerCAmelCase = InstructBlipVisionConfig(image_size=lowerCAmelCase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "t5-xl" in model_name: _lowerCAmelCase = TaConfig.from_pretrained("""google/flan-t5-xl""" , dense_act_fn="""gelu""" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: _lowerCAmelCase = TaConfig.from_pretrained("""google/flan-t5-xxl""" , dense_act_fn="""gelu""" , bos_token_id=1 ).to_dict() elif "vicuna-7b" in model_name: _lowerCAmelCase = LlamaConfig.from_pretrained("""decapoda-research/llama-7b-hf""" , vocab_size=3_20_01 ).to_dict() elif "vicuna-13b" in model_name: _lowerCAmelCase = LlamaConfig.from_pretrained("""decapoda-research/llama-13b-hf""" , vocab_size=3_20_01 ).to_dict() else: raise ValueError("""Model name not supported""" ) # the authors add one special "[DEC]" token to the vocab of Q-Former, hence vocab size = 30522 + 1 _lowerCAmelCase = InstructBlipQFormerConfig(vocab_size=3_05_23 ).to_dict() _lowerCAmelCase = InstructBlipConfig(vision_config=lowerCAmelCase , text_config=lowerCAmelCase , qformer_config=lowerCAmelCase ) return config, image_size @torch.no_grad() def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase=None , lowerCAmelCase=False ): """simple docstring""" _lowerCAmelCase = AutoTokenizer.from_pretrained("""bert-base-uncased""" , truncation_side="""left""" ) qformer_tokenizer.add_special_tokens({"""bos_token""": """[DEC]"""} ) if "t5" in model_name: _lowerCAmelCase = TaTokenizerFast.from_pretrained("""google/flan-t5-xl""" , truncation_side="""left""" ) elif "vicuna" in model_name: # the following was used in the original implementation: # tokenizer = LlamaTokenizer.from_pretrained("huggyllama/llama-7b", use_fast=False, truncation_side="left") # tokenizer.add_special_tokens({"pad_token": "[PAD]"}) # tokenizer.add_special_tokens({"bos_token": "</s>"}) # tokenizer.add_special_tokens({"eos_token": "</s>"}) # tokenizer.add_special_tokens({"unk_token": "</s>"}) _lowerCAmelCase = LlamaTokenizerFast.from_pretrained( """huggyllama/llama-7b""" , truncation_side="""left""" , bos_token="""</s>""" , unk_token="""</s>""" ) tokenizer.add_special_tokens({"""pad_token""": """[PAD]"""} ) _lowerCAmelCase , _lowerCAmelCase = get_blipa_config(lowerCAmelCase ) _lowerCAmelCase = InstructBlipForConditionalGeneration(lowerCAmelCase ).eval() _lowerCAmelCase = { """instructblip-vicuna-7b""": ("""blip2_vicuna_instruct""", """vicuna7b"""), """instructblip-vicuna-13b""": ("""blip2_vicuna_instruct""", """vicuna13b"""), """instructblip-flan-t5-xl""": ("""blip2_t5_instruct""", """flant5xl"""), """instructblip-flan-t5-xxl""": ("""blip2_t5_instruct""", """flant5xxl"""), } _lowerCAmelCase , _lowerCAmelCase = model_name_to_original[model_name] # load original model print("""Loading original model...""" ) _lowerCAmelCase = """cuda:1""" if torch.cuda.is_available() else """cpu""" _lowerCAmelCase = """cuda:2""" if torch.cuda.is_available() else """cpu""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = load_model_and_preprocess( name=lowerCAmelCase , model_type=lowerCAmelCase , is_eval=lowerCAmelCase , device=lowerCAmelCase ) original_model.eval() print("""Done!""" ) # update state dict keys _lowerCAmelCase = original_model.state_dict() _lowerCAmelCase = create_rename_keys(lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): _lowerCAmelCase = state_dict.pop(lowerCAmelCase ) if key.startswith("""Qformer.bert""" ): _lowerCAmelCase = key.replace("""Qformer.bert""" , """qformer""" ) if "attention.self" in key: _lowerCAmelCase = key.replace("""self""" , """attention""" ) if "llm_proj" in key: _lowerCAmelCase = key.replace("""llm_proj""" , """language_projection""" ) if "t5_proj" in key: _lowerCAmelCase = key.replace("""t5_proj""" , """language_projection""" ) if key.startswith("""llm_model""" ): _lowerCAmelCase = key.replace("""llm_model""" , """language_model""" ) if key.startswith("""t5""" ): _lowerCAmelCase = key.replace("""t5""" , """language""" ) _lowerCAmelCase = val # read in qv biases read_in_q_v_bias(lowerCAmelCase , lowerCAmelCase ) # note: weights get loaded in torch.float32 by default hf_model.load_state_dict(lowerCAmelCase , strict=lowerCAmelCase ) _lowerCAmelCase = load_demo_image() _lowerCAmelCase = """What is unusual about this image?""" # create processor _lowerCAmelCase = BlipImageProcessor( size={"""height""": image_size, """width""": image_size} , image_mean=lowerCAmelCase , image_std=lowerCAmelCase ) _lowerCAmelCase = InstructBlipProcessor( image_processor=lowerCAmelCase , tokenizer=lowerCAmelCase , qformer_tokenizer=lowerCAmelCase , ) _lowerCAmelCase = processor(images=lowerCAmelCase , text=lowerCAmelCase , return_tensors="""pt""" ).to(lowerCAmelCase ) # make sure processor creates exact same pixel values _lowerCAmelCase = vis_processors["""eval"""](lowerCAmelCase ).unsqueeze(0 ).to(lowerCAmelCase ) _lowerCAmelCase = inputs.pixel_values assert torch.allclose(original_pixel_values.to(pixel_values.device ) , lowerCAmelCase ) original_model.to(lowerCAmelCase ) hf_model.to(lowerCAmelCase ) with torch.no_grad(): if "vicuna" in model_name: _lowerCAmelCase = original_model({"""image""": original_pixel_values, """text_input""": [prompt]} ).logits _lowerCAmelCase = hf_model(**lowerCAmelCase ).logits else: _lowerCAmelCase = original_model( {"""image""": original_pixel_values, """text_input""": [prompt], """text_output""": ["""\n"""]} ).logits _lowerCAmelCase = tokenizer("""\n""" , return_tensors="""pt""" ).input_ids.to(lowerCAmelCase ) _lowerCAmelCase = label_input_ids.masked_fill(label_input_ids == tokenizer.pad_token_id , -1_00 ) _lowerCAmelCase = hf_model(**lowerCAmelCase , labels=lowerCAmelCase ).logits print("""First values of original logits:""" , original_logits[0, :3, :3] ) print("""First values of HF logits:""" , logits[0, :3, :3] ) # assert values assert original_logits.shape == logits.shape _lowerCAmelCase = 1e-4 if """vicuna""" in model_name else 1e-5 assert torch.allclose(original_logits.to(logits.device ) , lowerCAmelCase , atol=lowerCAmelCase ) print("""Looks ok!""" ) print("""Generating with original model...""" ) _lowerCAmelCase = original_model.generate({"""image""": original_pixel_values, """prompt""": prompt} , num_beams=5 ) # important: we need to cast the weights of the HF model to the appropriate type print("""Generating with HF model...""" ) _lowerCAmelCase = hf_model.generate( **lowerCAmelCase , do_sample=lowerCAmelCase , num_beams=5 , max_length=2_56 , min_length=1 , top_p=0.9 , repetition_penalty=1.5 , length_penalty=1.0 , temperature=1 , ) if "vicuna" in model_name: # convert output id 0 to 2 (eos_token_id) # TODO add this in the generate method? _lowerCAmelCase = 2 print("""Original generation:""" , lowerCAmelCase ) _lowerCAmelCase = processor.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) _lowerCAmelCase = [text.strip() for text in output_text] print("""HF generation:""" , lowerCAmelCase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(lowerCAmelCase ) hf_model.save_pretrained(lowerCAmelCase ) if push_to_hub: processor.push_to_hub(f"Salesforce/{model_name}" ) hf_model.push_to_hub(f"Salesforce/{model_name}" ) if __name__ == "__main__": A__ : List[str] =argparse.ArgumentParser() A__ : Union[str, Any] =[ '''instructblip-vicuna-7b''', '''instructblip-vicuna-13b''', '''instructblip-flan-t5-xl''', '''instructblip-flan-t5-xxl''', ] parser.add_argument( '''--model_name''', default='''instructblip-flan-t5-xl''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) A__ : Any =parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
220
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets A__ : List[str] =datasets.logging.get_logger(__name__) A__ : List[Any] ='''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' A__ : List[str] ='''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' A__ : List[Any] =''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase="dummy_doc" ): """simple docstring""" _lowerCAmelCase = {doc: key_lines} _lowerCAmelCase = {doc: sys_lines} _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase , sys_doc_lines[doc] , lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase , lowerCAmelCase ) if remove_nested: _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase , lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase , lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ f"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ f"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( f"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_coref_infos(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 for name, metric in metrics: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = evaluator.evaluate_documents(lowerCAmelCase , lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"{name}/recall": recall, f"{name}/precision": precision, f"{name}/f1": fa} ) logger.info( name.ljust(10 ) , f"Recall: {recall * 1_00:.2f}" , f" Precision: {precision * 1_00:.2f}" , f" F1: {fa * 1_00:.2f}" , ) if conll_subparts_num == 3: _lowerCAmelCase = (conll / 3) * 1_00 logger.info(f"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: _lowerCAmelCase = line.split()[5] if not parse_col == "-": _lowerCAmelCase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): def lowercase__ ( self : str ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def lowercase__ ( self : List[Any] , __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=True , __snake_case : List[str]=False , __snake_case : List[Any]=False , __snake_case : Dict=False ) -> Union[str, Any]: _lowerCAmelCase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _lowerCAmelCase = util.check_gold_parse_annotation(__snake_case ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _lowerCAmelCase = evaluate( key_lines=__snake_case , sys_lines=__snake_case , metrics=__snake_case , NP_only=__snake_case , remove_nested=__snake_case , keep_singletons=__snake_case , min_span=__snake_case , ) return score
220
1
import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging A__ = logging.get_logger(__name__) A__ = { """facebook/encodec_24khz""": """https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json""", """facebook/encodec_48khz""": """https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCamelCase__ ): __lowerCamelCase = '''encodec''' def __init__( self , _snake_case=[1.5, 3.0, 6.0, 12.0, 24.0] , _snake_case=24000 , _snake_case=1 , _snake_case=False , _snake_case=None , _snake_case=None , _snake_case=128 , _snake_case=32 , _snake_case=1 , _snake_case=[8, 5, 4, 2] , _snake_case="weight_norm" , _snake_case=7 , _snake_case=7 , _snake_case=3 , _snake_case=2 , _snake_case=True , _snake_case="reflect" , _snake_case=2 , _snake_case=2 , _snake_case=1.0 , _snake_case=1024 , _snake_case=None , _snake_case=True , **_snake_case , ): """simple docstring""" _lowerCAmelCase = target_bandwidths _lowerCAmelCase = sampling_rate _lowerCAmelCase = audio_channels _lowerCAmelCase = normalize _lowerCAmelCase = chunk_length_s _lowerCAmelCase = overlap _lowerCAmelCase = hidden_size _lowerCAmelCase = num_filters _lowerCAmelCase = num_residual_layers _lowerCAmelCase = upsampling_ratios _lowerCAmelCase = norm_type _lowerCAmelCase = kernel_size _lowerCAmelCase = last_kernel_size _lowerCAmelCase = residual_kernel_size _lowerCAmelCase = dilation_growth_rate _lowerCAmelCase = use_causal_conv _lowerCAmelCase = pad_mode _lowerCAmelCase = compress _lowerCAmelCase = num_lstm_layers _lowerCAmelCase = trim_right_ratio _lowerCAmelCase = codebook_size _lowerCAmelCase = codebook_dim if codebook_dim is not None else hidden_size _lowerCAmelCase = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**_snake_case ) @property def snake_case ( self ): """simple docstring""" if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def snake_case ( self ): """simple docstring""" if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def snake_case ( self ): """simple docstring""" return int(1000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
82
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class __lowerCAmelCase ( unittest.TestCase ): def snake_case ( self ): """simple docstring""" _lowerCAmelCase = { """task_specific_params""": { """summarization""": {"""length_penalty""": 1.0, """max_length""": 128, """min_length""": 12, """num_beams""": 4}, """summarization_cnn""": {"""length_penalty""": 2.0, """max_length""": 142, """min_length""": 56, """num_beams""": 4}, """summarization_xsum""": {"""length_penalty""": 1.0, """max_length""": 62, """min_length""": 11, """num_beams""": 6}, } } _lowerCAmelCase = { """task_specific_params.summarization.length_penalty""": 1.0, """task_specific_params.summarization.max_length""": 128, """task_specific_params.summarization.min_length""": 12, """task_specific_params.summarization.num_beams""": 4, """task_specific_params.summarization_cnn.length_penalty""": 2.0, """task_specific_params.summarization_cnn.max_length""": 142, """task_specific_params.summarization_cnn.min_length""": 56, """task_specific_params.summarization_cnn.num_beams""": 4, """task_specific_params.summarization_xsum.length_penalty""": 1.0, """task_specific_params.summarization_xsum.max_length""": 62, """task_specific_params.summarization_xsum.min_length""": 11, """task_specific_params.summarization_xsum.num_beams""": 6, } self.assertEqual(flatten_dict(_snake_case ) , _snake_case ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(_snake_case ) , x.transpose() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , transpose(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , transpose(_snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , transpose(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , transpose(_snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case ) , np.asarray(transpose(_snake_case ) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(transpose(_snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(_snake_case , axes=(1, 2, 0) ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , np.reshape(_snake_case , (4, 3) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , np.reshape(_snake_case , (12, 5) ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , reshape(_snake_case , (4, 3) ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , reshape(_snake_case , (12, 5) ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , reshape(_snake_case , (4, 3) ).numpy() ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , reshape(_snake_case , (12, 5) ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (4, 3) ) , np.asarray(reshape(_snake_case , (4, 3) ) ) ) ) _lowerCAmelCase = np.random.randn(3 , 4 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(reshape(_snake_case , (12, 5) ) , np.asarray(reshape(_snake_case , (12, 5) ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(_snake_case ) , np.squeeze(_snake_case ) ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , np.squeeze(_snake_case , axis=2 ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , squeeze(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , squeeze(_snake_case , axis=2 ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , squeeze(_snake_case ).numpy() ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , squeeze(_snake_case , axis=2 ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(1 , 3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case ) , np.asarray(squeeze(_snake_case ) ) ) ) _lowerCAmelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(squeeze(_snake_case , axis=2 ) , np.asarray(squeeze(_snake_case , axis=2 ) ) ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , np.expand_dims(_snake_case , axis=1 ) ) ) @require_torch def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = torch.tensor(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , expand_dims(_snake_case , axis=1 ).numpy() ) ) @require_tf def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = tf.constant(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , expand_dims(_snake_case , axis=1 ).numpy() ) ) @require_flax def snake_case ( self ): """simple docstring""" _lowerCAmelCase = np.random.randn(3 , 4 ) _lowerCAmelCase = jnp.array(_snake_case ) self.assertTrue(np.allclose(expand_dims(_snake_case , axis=1 ) , np.asarray(expand_dims(_snake_case , axis=1 ) ) ) )
82
1
import string def lowerCAmelCase_ ( UpperCamelCase_ ) -> None: for key in range(len(string.ascii_uppercase ) ): UpperCamelCase_ = '''''' for symbol in message: if symbol in string.ascii_uppercase: UpperCamelCase_ = string.ascii_uppercase.find(UpperCamelCase_ ) UpperCamelCase_ = num - key if num < 0: UpperCamelCase_ = num + len(string.ascii_uppercase ) UpperCamelCase_ = translated + string.ascii_uppercase[num] else: UpperCamelCase_ = translated + symbol print(F'''Decryption using Key #{key}: {translated}''' ) def lowerCAmelCase_ ( ) -> None: UpperCamelCase_ = input("Encrypted message: " ) UpperCamelCase_ = message.upper() decrypt(UpperCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
361
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys _UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
328
0
"""simple docstring""" from __future__ import annotations from random import random class __A : '''simple docstring''' def __init__( self : str ,_snake_case : int | None = None ) -> Any: """simple docstring""" lowercase__ : int = value lowercase__ : Any = random() lowercase__ : Node | None = None lowercase__ : Node | None = None def __repr__( self : Tuple ) -> str: """simple docstring""" from pprint import pformat if self.left is None and self.right is None: return f"""'{self.value}: {self.prior:.5}'""" else: return pformat( {f"""{self.value}: {self.prior:.5}""": (self.left, self.right)} ,indent=1 ) def __str__( self : Dict ) -> str: """simple docstring""" lowercase__ : Dict = str(self.value ) + ''' ''' lowercase__ : Dict = str(self.left or '''''' ) lowercase__ : List[str] = str(self.right or '''''' ) return value + left + right def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> tuple[Node | None, Node | None]: if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase__ , lowercase__ : Dict = split(root.left , __lowerCamelCase ) return left, root else: lowercase__ , lowercase__ : Optional[Any] = split(root.right , __lowerCamelCase ) return root, right def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> Node | None: if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase__ : Optional[int] = merge(left.right , __lowerCamelCase ) return left else: lowercase__ : str = merge(__lowerCamelCase , right.left ) return right def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> Node | None: lowercase__ : Union[str, Any] = Node(__lowerCamelCase ) lowercase__ , lowercase__ : Optional[Any] = split(__lowerCamelCase , __lowerCamelCase ) return merge(merge(__lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase ) def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> Node | None: lowercase__ , lowercase__ : List[str] = split(__lowerCamelCase , value - 1 ) lowercase__ , lowercase__ : Tuple = split(__lowerCamelCase , __lowerCamelCase ) return merge(__lowerCamelCase , __lowerCamelCase ) def __UpperCAmelCase ( __lowerCamelCase ) -> None: if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> Node | None: for arg in args.split(): if arg[0] == "+": lowercase__ : Any = insert(__lowerCamelCase , int(arg[1:] ) ) elif arg[0] == "-": lowercase__ : List[Any] = erase(__lowerCamelCase , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def __UpperCAmelCase ( ) -> None: lowercase__ : str = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase__ : Optional[Any] = input() while args != "q": lowercase__ : Union[str, Any] = interact_treap(__lowerCamelCase , __lowerCamelCase ) print(__lowerCamelCase ) lowercase__ : Optional[Any] = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
16
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __A : '''simple docstring''' def __init__( self : str ,_snake_case : List[Any] ,_snake_case : Optional[int]=3 ,_snake_case : Optional[int]=32 ,_snake_case : Union[str, Any]=3 ,_snake_case : int=10 ,_snake_case : List[str]=[10, 20, 30, 40] ,_snake_case : Any=[1, 1, 2, 1] ,_snake_case : int=True ,_snake_case : Optional[Any]=True ,_snake_case : Union[str, Any]="relu" ,_snake_case : Dict=3 ,_snake_case : Any=None ,) -> str: """simple docstring""" lowercase__ : int = parent lowercase__ : Optional[Any] = batch_size lowercase__ : Optional[Any] = image_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = embeddings_size lowercase__ : Optional[Any] = hidden_sizes lowercase__ : str = depths lowercase__ : Tuple = is_training lowercase__ : List[Any] = use_labels lowercase__ : Union[str, Any] = hidden_act lowercase__ : Union[str, Any] = num_labels lowercase__ : Tuple = scope lowercase__ : Optional[Any] = len(_snake_case ) def UpperCAmelCase ( self : Optional[int] ) -> Tuple: """simple docstring""" lowercase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ : Tuple = None if self.use_labels: lowercase__ : Dict = ids_tensor([self.batch_size] ,self.num_labels ) lowercase__ : int = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" return ResNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def UpperCAmelCase ( self : List[str] ,_snake_case : Optional[int] ,_snake_case : int ,_snake_case : Tuple ) -> List[Any]: """simple docstring""" lowercase__ : Optional[int] = TFResNetModel(config=_snake_case ) lowercase__ : List[str] = model(_snake_case ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def UpperCAmelCase ( self : Optional[int] ,_snake_case : Optional[Any] ,_snake_case : int ,_snake_case : Any ) -> Tuple: """simple docstring""" lowercase__ : Tuple = self.num_labels lowercase__ : Union[str, Any] = TFResNetForImageClassification(_snake_case ) lowercase__ : List[str] = model(_snake_case ,labels=_snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCAmelCase ( self : Tuple ) -> str: """simple docstring""" lowercase__ : Dict = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = config_and_inputs lowercase__ : Dict = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class __A ( A_ ,A_ ,unittest.TestCase ): '''simple docstring''' lowerCAmelCase : Optional[int] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () lowerCAmelCase : Any = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) lowerCAmelCase : List[Any] = False lowerCAmelCase : List[Any] = False lowerCAmelCase : int = False lowerCAmelCase : Union[str, Any] = False lowerCAmelCase : List[str] = False def UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowercase__ : Optional[Any] = TFResNetModelTester(self ) lowercase__ : int = ConfigTester(self ,config_class=_snake_case ,has_text_modality=_snake_case ) def UpperCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self : List[Any] ) -> List[str]: """simple docstring""" return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def UpperCAmelCase ( self : Optional[int] ) -> Dict: """simple docstring""" pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def UpperCAmelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def UpperCAmelCase ( self : int ) -> Union[str, Any]: """simple docstring""" lowercase__ , lowercase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ : str = model_class(_snake_case ) lowercase__ : Dict = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ : Optional[int] = [*signature.parameters.keys()] lowercase__ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,_snake_case ) def UpperCAmelCase ( self : Tuple ) -> Any: """simple docstring""" lowercase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCAmelCase ( self : Dict ) -> List[str]: """simple docstring""" def check_hidden_states_output(_snake_case : Optional[int] ,_snake_case : List[str] ,_snake_case : Optional[Any] ): lowercase__ : str = model_class(_snake_case ) lowercase__ : Union[str, Any] = model(**self._prepare_for_class(_snake_case ,_snake_case ) ) lowercase__ : List[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase__ : Tuple = self.model_tester.num_stages self.assertEqual(len(_snake_case ) ,expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 4, self.model_tester.image_size // 4] ,) lowercase__ , lowercase__ : int = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : List[Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase__ : List[Any] = layer_type lowercase__ : Dict = True check_hidden_states_output(_snake_case ,_snake_case ,_snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ : Dict = True check_hidden_states_output(_snake_case ,_snake_case ,_snake_case ) def UpperCAmelCase ( self : Dict ) -> Any: """simple docstring""" lowercase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_snake_case ) @slow def UpperCAmelCase ( self : Optional[Any] ) -> int: """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Optional[Any] = TFResNetModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def __UpperCAmelCase ( ) -> Dict: lowercase__ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class __A ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCAmelCase ( self : str ) -> Any: """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase ( self : Union[str, Any] ) -> Dict: """simple docstring""" lowercase__ : Tuple = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase__ : Any = self.default_image_processor lowercase__ : int = prepare_img() lowercase__ : Tuple = image_processor(images=_snake_case ,return_tensors='''tf''' ) # forward pass lowercase__ : Dict = model(**_snake_case ) # verify the logits lowercase__ : List[str] = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape ,_snake_case ) lowercase__ : Any = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() ,_snake_case ,atol=1e-4 ) )
16
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase = """ Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\") >>> repo = \"openai/shap-e-img2img\" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\" >>> image = load_image(image_url).convert(\"RGB\") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\") ``` """ @dataclass class lowerCAmelCase_ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case = 42 class lowerCAmelCase_ ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): super().__init__() self.register_modules( prior=__lowerCAmelCase , image_encoder=__lowerCAmelCase , image_processor=__lowerCAmelCase , scheduler=__lowerCAmelCase , renderer=__lowerCAmelCase , ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): if latents is None: snake_case_ = randn_tensor(__lowerCAmelCase , generator=__lowerCAmelCase , device=__lowerCAmelCase , dtype=__lowerCAmelCase ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) snake_case_ = latents.to(__lowerCAmelCase ) snake_case_ = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , _UpperCAmelCase=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case_ = torch.device(F'''cuda:{gpu_id}''' ) snake_case_ = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__lowerCAmelCase , __lowerCAmelCase ) @property def UpperCamelCase__ ( self ): if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder , '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(__lowerCAmelCase , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): if isinstance(__lowerCAmelCase , __lowerCAmelCase ) and isinstance(image[0] , torch.Tensor ): snake_case_ = torch.cat(__lowerCAmelCase , axis=0 ) if image[0].ndim == 4 else torch.stack(__lowerCAmelCase , axis=0 ) if not isinstance(__lowerCAmelCase , torch.Tensor ): snake_case_ = self.image_processor(__lowerCAmelCase , return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) snake_case_ = image.to(dtype=self.image_encoder.dtype , device=__lowerCAmelCase ) snake_case_ = self.image_encoder(__lowerCAmelCase )['''last_hidden_state'''] snake_case_ = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 snake_case_ = image_embeds.repeat_interleave(__lowerCAmelCase , dim=0 ) if do_classifier_free_guidance: snake_case_ = torch.zeros_like(__lowerCAmelCase ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(__lowerCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase = 1 , _UpperCAmelCase = 25 , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = 4.0 , _UpperCAmelCase = 64 , _UpperCAmelCase = "pil" , _UpperCAmelCase = True , ): if isinstance(__lowerCAmelCase , PIL.Image.Image ): snake_case_ = 1 elif isinstance(__lowerCAmelCase , torch.Tensor ): snake_case_ = image.shape[0] elif isinstance(__lowerCAmelCase , __lowerCAmelCase ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): snake_case_ = len(__lowerCAmelCase ) else: raise ValueError( F'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(__lowerCAmelCase )}''' ) snake_case_ = self._execution_device snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_image(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # prior self.scheduler.set_timesteps(__lowerCAmelCase , device=__lowerCAmelCase ) snake_case_ = self.scheduler.timesteps snake_case_ = self.prior.config.num_embeddings snake_case_ = self.prior.config.embedding_dim snake_case_ = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim snake_case_ = latents.reshape(latents.shape[0] , __lowerCAmelCase , __lowerCAmelCase ) for i, t in enumerate(self.progress_bar(__lowerCAmelCase ) ): # expand the latents if we are doing classifier free guidance snake_case_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case_ = self.scheduler.scale_model_input(__lowerCAmelCase , __lowerCAmelCase ) snake_case_ = self.prior( __lowerCAmelCase , timestep=__lowerCAmelCase , proj_embedding=__lowerCAmelCase , ).predicted_image_embedding # remove the variance snake_case_ , snake_case_ = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: snake_case_ , snake_case_ = noise_pred.chunk(2 ) snake_case_ = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) snake_case_ = self.scheduler.step( __lowerCAmelCase , timestep=__lowerCAmelCase , sample=__lowerCAmelCase , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=__lowerCAmelCase ) snake_case_ = [] for i, latent in enumerate(__lowerCAmelCase ): print() snake_case_ = self.renderer.decode( latent[None, :] , __lowerCAmelCase , size=__lowerCAmelCase , ray_batch_size=40_96 , n_coarse_samples=64 , n_fine_samples=1_28 , ) images.append(__lowerCAmelCase ) snake_case_ = torch.stack(__lowerCAmelCase ) if output_type not in ["np", "pil"]: raise ValueError(F'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) snake_case_ = images.cpu().numpy() if output_type == "pil": snake_case_ = [self.numpy_to_pil(__lowerCAmelCase ) for image in images] # Offload last model to CPU if hasattr(self , '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=__lowerCAmelCase )
358
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase = get_tests_dir("""fixtures/test_sentencepiece_no_bos.model""") @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' __snake_case = PegasusTokenizer __snake_case = PegasusTokenizerFast __snake_case = True __snake_case = True def UpperCamelCase__ ( self ): super().setUp() # We have a SentencePiece fixture for testing snake_case_ = PegasusTokenizer(_UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__ ( self ): return PegasusTokenizer.from_pretrained('''google/pegasus-large''' ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase ): return ("This is a test", "This is a test") def UpperCamelCase__ ( self ): snake_case_ = '''</s>''' snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''</s>''' ) self.assertEqual(vocab_keys[-1] , '''v''' ) self.assertEqual(len(_UpperCAmelCase ) , 11_03 ) def UpperCamelCase__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 11_03 ) def UpperCamelCase__ ( self ): snake_case_ = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = self.tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = ( '''Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important''' ''' </s> <pad> <pad> <pad>''' ) snake_case_ = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] snake_case_ = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word snake_case_ = '''<mask_1> To ensure a <mask_2> flow of bank resolutions.''' snake_case_ = [2, 4_13, 6_15, 1_14, 3, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] snake_case_ = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_61_03 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_03 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_05 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 10_24 snake_case_ = '''To ensure a smooth flow of bank resolutions.''' snake_case_ = [4_13, 6_15, 1_14, 22_91, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] snake_case_ = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def UpperCamelCase__ ( self ): snake_case_ = ['''This is going to be way too long.''' * 1_50, '''short example'''] snake_case_ = ['''not super long but more than 5 tokens''', '''tiny'''] snake_case_ = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) snake_case_ = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 10_24) assert batch.attention_mask.shape == (2, 10_24) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. @slow def UpperCamelCase__ ( self ): # fmt: off snake_case_ = {'''input_ids''': [[3_89_79, 1_43, 1_84_85, 6_06, 1_30, 2_66_69, 8_76_86, 1_21, 5_41_89, 11_29, 1_11, 2_66_69, 8_76_86, 1_21, 91_14, 1_47_87, 1_21, 1_32_49, 1_58, 5_92, 9_56, 1_21, 1_46_21, 3_15_76, 1_43, 6_26_13, 1_08, 96_88, 9_30, 4_34_30, 1_15_62, 6_26_13, 3_04, 1_08, 1_14_43, 8_97, 1_08, 93_14, 1_74_15, 6_33_99, 1_08, 1_14_43, 76_14, 1_83_16, 1_18, 42_84, 71_48, 1_24_30, 1_43, 14_00, 2_57_03, 1_58, 1_11, 42_84, 71_48, 1_17_72, 1_43, 2_12_97, 10_64, 1_58, 1_22, 2_04, 35_06, 17_54, 11_33, 1_47_87, 15_81, 1_15, 3_32_24, 44_82, 1_11, 13_55, 1_10, 2_91_73, 3_17, 5_08_33, 1_08, 2_01_47, 9_46_65, 1_11, 7_71_98, 1_07, 1], [1_10, 6_26_13, 1_17, 6_38, 1_12, 11_33, 1_21, 2_00_98, 13_55, 7_90_50, 1_38_72, 1_35, 15_96, 5_35_41, 13_52, 1_41, 1_30_39, 55_42, 1_24, 3_02, 5_18, 1_11, 2_68, 29_56, 1_15, 1_49, 44_27, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_39, 12_35, 27_99, 1_82_89, 1_77_80, 2_04, 1_09, 94_74, 12_96, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='''google/bigbird-pegasus-large-arxiv''' , revision='''ba85d0851d708441f91440d509690f1ab6353415''' , ) @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' __snake_case = PegasusTokenizer __snake_case = PegasusTokenizerFast __snake_case = True __snake_case = True def UpperCamelCase__ ( self ): super().setUp() # We have a SentencePiece fixture for testing snake_case_ = PegasusTokenizer(_UpperCAmelCase , offset=0 , mask_token_sent=_UpperCAmelCase , mask_token='''[MASK]''' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__ ( self ): return PegasusTokenizer.from_pretrained('''google/bigbird-pegasus-large-arxiv''' ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase ): return ("This is a test", "This is a test") def UpperCamelCase__ ( self ): snake_case_ = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = self.tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = ( '''Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>''' ''' <pad> <pad> <pad>''' ) snake_case_ = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] snake_case_ = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @require_torch def UpperCamelCase__ ( self ): snake_case_ = ['''This is going to be way too long.''' * 10_00, '''short example'''] snake_case_ = ['''not super long but more than 5 tokens''', '''tiny'''] snake_case_ = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) snake_case_ = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 40_96) assert batch.attention_mask.shape == (2, 40_96) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. def UpperCamelCase__ ( self ): snake_case_ = ( '''This is an example string that is used to test the original TF implementation against the HF''' ''' implementation''' ) snake_case_ = self._large_tokenizer(_UpperCAmelCase ).input_ids self.assertListEqual( _UpperCAmelCase , [1_82, 1_17, 1_42, 5_87, 42_11, 1_20, 1_17, 2_63, 1_12, 8_04, 1_09, 8_56, 2_50_16, 31_37, 4_64, 1_09, 2_69_55, 31_37, 1] , )
267
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin A : Any = get_tests_dir("fixtures/test_sentencepiece.model") A : List[str] = {"target_lang": "fi", "source_lang": "en"} A : Dict = ">>zh<<" A : Union[str, Any] = "Helsinki-NLP/" if is_torch_available(): A : Optional[int] = "pt" elif is_tf_available(): A : str = "tf" else: A : str = "jax" @require_sentencepiece class _lowercase ( lowercase__ , unittest.TestCase): """simple docstring""" A__ = MarianTokenizer A__ = False A__ = True def lowerCAmelCase ( self : Dict ): '''simple docstring''' super().setUp() lowerCamelCase__ : Optional[Any] = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] lowerCamelCase__ : str = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) lowerCamelCase__ : List[str] = Path(self.tmpdirname ) save_json(__lowerCamelCase , save_dir / VOCAB_FILES_NAMES["vocab"] ) save_json(__lowerCamelCase , save_dir / VOCAB_FILES_NAMES["tokenizer_config_file"] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(__lowerCamelCase , save_dir / VOCAB_FILES_NAMES["source_spm"] ) copyfile(__lowerCamelCase , save_dir / VOCAB_FILES_NAMES["target_spm"] ) lowerCamelCase__ : Tuple = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase ( self : Optional[int] , **__lowerCamelCase : Tuple ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowerCAmelCase ( self : Any , __lowerCamelCase : List[Any] ): '''simple docstring''' return ( "This is a test", "This is a test", ) def lowerCAmelCase ( self : Dict ): '''simple docstring''' lowerCamelCase__ : str = "</s>" lowerCamelCase__ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' lowerCamelCase__ : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<pad>" ) self.assertEqual(len(__lowerCamelCase ) , 9 ) def lowerCAmelCase ( self : int ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 9 ) def lowerCAmelCase ( self : Tuple ): '''simple docstring''' lowerCamelCase__ : Union[str, Any] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) lowerCamelCase__ : Any = en_de_tokenizer(["I am a small frog"] , return_tensors=__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : str = [38, 121, 14, 697, 38848, 0] self.assertListEqual(__lowerCamelCase , batch.input_ids[0] ) lowerCamelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(__lowerCamelCase ) lowerCamelCase__ : int = [x.name for x in Path(__lowerCamelCase ).glob("*" )] self.assertIn("source.spm" , __lowerCamelCase ) MarianTokenizer.from_pretrained(__lowerCamelCase ) def lowerCAmelCase ( self : List[str] ): '''simple docstring''' lowerCamelCase__ : str = self.get_tokenizer() lowerCamelCase__ : Any = tok( ["I am a small frog" * 1000, "I am a small frog"] , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors=__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertEqual(batch.input_ids.shape , (2, 512) ) def lowerCAmelCase ( self : Tuple ): '''simple docstring''' lowerCamelCase__ : Optional[int] = self.get_tokenizer() lowerCamelCase__ : Union[str, Any] = tok(["I am a tiny frog", "I am a small frog"] , padding=__lowerCamelCase , return_tensors=__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertEqual(batch_smaller.input_ids.shape , (2, 10) ) @slow def lowerCAmelCase ( self : Optional[Any] ): '''simple docstring''' lowerCamelCase__ : Optional[Any] = {"input_ids": [[43495, 462, 20, 42164, 1369, 52, 464, 132, 1703, 492, 13, 7491, 38999, 6, 8, 464, 132, 1703, 492, 13, 4669, 37867, 13, 7525, 27, 1593, 988, 13, 33972, 7029, 6, 20, 8251, 383, 2, 270, 5866, 3788, 2, 2353, 8251, 12338, 2, 13958, 387, 2, 3629, 6953, 188, 2900, 2, 13958, 8011, 11501, 23, 8460, 4073, 34009, 20, 435, 11439, 27, 8, 8460, 4073, 6004, 20, 9988, 375, 27, 33, 266, 1945, 1076, 1350, 37867, 3288, 5, 577, 1076, 4374, 8, 5082, 5, 26453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10767, 6, 316, 304, 4239, 3, 0], [148, 15722, 19, 1839, 12, 1350, 13, 22327, 5082, 5418, 47567, 35938, 59, 318, 19552, 108, 2183, 54, 14976, 4835, 32, 547, 1114, 8, 315, 2417, 5, 92, 19088, 3, 0, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100], [36, 6395, 12570, 39147, 11597, 6, 266, 4, 45405, 7296, 3, 0, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100, 58100]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="Helsinki-NLP/opus-mt-en-de" , revision="1a8c2263da11e68e50938f97e10cd57820bd504c" , decode_kwargs={"use_source_tokenizer": True} , ) def lowerCAmelCase ( self : Any ): '''simple docstring''' lowerCamelCase__ : Tuple = MarianTokenizer.from_pretrained("hf-internal-testing/test-marian-two-vocabs" ) lowerCamelCase__ : Any = "Tämä on testi" lowerCamelCase__ : Optional[int] = "This is a test" lowerCamelCase__ : Optional[Any] = [76, 7, 2047, 2] lowerCamelCase__ : List[str] = [69, 12, 11, 940, 2] lowerCamelCase__ : Optional[Any] = tokenizer(__lowerCamelCase ).input_ids self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : List[str] = tokenizer(text_target=__lowerCamelCase ).input_ids self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , __lowerCamelCase )
184
from collections import defaultdict def lowercase_ ( _A : int ): """simple docstring""" lowerCamelCase__ : Union[str, Any] = 1 lowerCamelCase__ : Dict = True for v in tree[start]: if v not in visited: ret += dfs(_A ) if ret % 2 == 0: cuts.append(_A ) return ret def lowercase_ ( ): """simple docstring""" dfs(1 ) if __name__ == "__main__": A, A : Tuple = 10, 9 A : int = defaultdict(list) A : dict[int, bool] = {} A : list[int] = [] A : List[str] = 0 A : Tuple = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
184
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase : Optional[int] = { """configuration_distilbert""": [ """DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DistilBertConfig""", """DistilBertOnnxConfig""", ], """tokenization_distilbert""": ["""DistilBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : Dict = ["""DistilBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : Optional[Any] = [ """DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """DistilBertForMaskedLM""", """DistilBertForMultipleChoice""", """DistilBertForQuestionAnswering""", """DistilBertForSequenceClassification""", """DistilBertForTokenClassification""", """DistilBertModel""", """DistilBertPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : List[Any] = [ """TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDistilBertForMaskedLM""", """TFDistilBertForMultipleChoice""", """TFDistilBertForQuestionAnswering""", """TFDistilBertForSequenceClassification""", """TFDistilBertForTokenClassification""", """TFDistilBertMainLayer""", """TFDistilBertModel""", """TFDistilBertPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : Union[str, Any] = [ """FlaxDistilBertForMaskedLM""", """FlaxDistilBertForMultipleChoice""", """FlaxDistilBertForQuestionAnswering""", """FlaxDistilBertForSequenceClassification""", """FlaxDistilBertForTokenClassification""", """FlaxDistilBertModel""", """FlaxDistilBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys UpperCamelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
364
'''simple docstring''' import logging from transformers.configuration_utils import PretrainedConfig UpperCamelCase : Optional[Any] = logging.getLogger(__name__) class UpperCamelCase ( a_ ): """simple docstring""" A : Tuple = "masked_bert" def __init__( self : Tuple , UpperCAmelCase_ : List[Any]=3_0_5_2_2 , UpperCAmelCase_ : str=7_6_8 , UpperCAmelCase_ : Optional[Any]=1_2 , UpperCAmelCase_ : Optional[int]=1_2 , UpperCAmelCase_ : Union[str, Any]=3_0_7_2 , UpperCAmelCase_ : Union[str, Any]="gelu" , UpperCAmelCase_ : Optional[Any]=0.1 , UpperCAmelCase_ : List[Any]=0.1 , UpperCAmelCase_ : Optional[int]=5_1_2 , UpperCAmelCase_ : List[str]=2 , UpperCAmelCase_ : str=0.02 , UpperCAmelCase_ : Optional[Any]=1e-12 , UpperCAmelCase_ : Dict=0 , UpperCAmelCase_ : Dict="topK" , UpperCAmelCase_ : str="constant" , UpperCAmelCase_ : Optional[Any]=0.0 , **UpperCAmelCase_ : Optional[Any] , ): """simple docstring""" super().__init__(pad_token_id=UpperCAmelCase_ , **UpperCAmelCase_) a : Union[str, Any] = vocab_size a : List[Any] = hidden_size a : List[str] = num_hidden_layers a : Any = num_attention_heads a : Optional[Any] = hidden_act a : str = intermediate_size a : Dict = hidden_dropout_prob a : Any = attention_probs_dropout_prob a : Any = max_position_embeddings a : Dict = type_vocab_size a : List[str] = initializer_range a : int = layer_norm_eps a : Dict = pruning_method a : List[str] = mask_init a : Union[str, Any] = mask_scale
345
0
'''simple docstring''' import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) __snake_case =logging.getLogger(__name__) @dataclass(frozen=__lowercase ) class UpperCAmelCase_ : lowerCamelCase : str lowerCamelCase : str lowerCamelCase : Optional[str] = None lowerCamelCase : Optional[str] = None lowerCamelCase : Optional[str] = None @dataclass(frozen=__lowercase ) class UpperCAmelCase_ : lowerCamelCase : List[int] lowerCamelCase : Optional[List[int]] = None lowerCamelCase : Optional[List[int]] = None lowerCamelCase : Optional[Union[int, float]] = None lowerCamelCase : Optional[int] = None if is_torch_available(): import torch from torch.utils.data import Dataset class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : List[InputFeatures] def __init__( self : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : PreTrainedTokenizer , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : bool = False , ) -> int: lowerCAmelCase = hans_processors[task]() lowerCAmelCase = os.path.join( UpperCAmelCase__ , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase__ ) , UpperCAmelCase__ , ) , ) lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase = cached_features_file + '.lock' with FileLock(UpperCAmelCase__ ): if os.path.exists(UpperCAmelCase__ ) and not overwrite_cache: logger.info(F'''Loading features from cached file {cached_features_file}''' ) lowerCAmelCase = torch.load(UpperCAmelCase__ ) else: logger.info(F'''Creating features from dataset file at {data_dir}''' ) lowerCAmelCase = ( processor.get_dev_examples(UpperCAmelCase__ ) if evaluate else processor.get_train_examples(UpperCAmelCase__ ) ) logger.info('Training examples: %s' , len(UpperCAmelCase__ ) ) lowerCAmelCase = hans_convert_examples_to_features(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) logger.info('Saving features into cached file %s' , UpperCAmelCase__ ) torch.save(self.features , UpperCAmelCase__ ) def __len__( self : List[str] ) -> Optional[int]: return len(self.features ) def __getitem__( self : Optional[int] , UpperCAmelCase__ : Optional[int] ) -> InputFeatures: return self.features[i] def __UpperCAmelCase ( self : Any ) -> Any: return self.label_list if is_tf_available(): import tensorflow as tf class UpperCAmelCase_ : lowerCamelCase : List[InputFeatures] def __init__( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : PreTrainedTokenizer , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] = 1_2_8 , UpperCAmelCase__ : str=False , UpperCAmelCase__ : bool = False , ) -> int: lowerCAmelCase = hans_processors[task]() lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list lowerCAmelCase = processor.get_dev_examples(UpperCAmelCase__ ) if evaluate else processor.get_train_examples(UpperCAmelCase__ ) lowerCAmelCase = hans_convert_examples_to_features(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 1_0_0_0_0 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase__ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCAmelCase = tf.data.Dataset.from_generator( UpperCAmelCase__ , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def __UpperCAmelCase ( self : Dict ) -> Any: return self.dataset def __len__( self : List[Any] ) -> List[str]: return len(self.features ) def __getitem__( self : int , UpperCAmelCase__ : int ) -> InputFeatures: return self.features[i] def __UpperCAmelCase ( self : Tuple ) -> int: return self.label_list class UpperCAmelCase_ ( __lowercase ): def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : Dict ) -> Any: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase__ , 'heuristics_train_set.txt' ) ) , 'train' ) def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : List[Any] ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase__ , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def __UpperCAmelCase ( self : List[str] ) -> Any: return ["contradiction", "entailment", "neutral"] def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] ) -> Tuple: lowerCAmelCase = [] for i, line in enumerate(UpperCAmelCase__ ): if i == 0: continue lowerCAmelCase = '%s-%s' % (set_type, line[0]) lowerCAmelCase = line[5] lowerCAmelCase = line[6] lowerCAmelCase = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCAmelCase = line[0] examples.append(InputExample(guid=UpperCAmelCase__ , text_a=UpperCAmelCase__ , text_b=UpperCAmelCase__ , label=UpperCAmelCase__ , pairID=UpperCAmelCase__ ) ) return examples def a_ ( lowerCamelCase : List[InputExample] , lowerCamelCase : List[str] , lowerCamelCase : int , lowerCamelCase : PreTrainedTokenizer , ): lowerCAmelCase = {label: i for i, label in enumerate(lowerCamelCase )} lowerCAmelCase = [] for ex_index, example in tqdm.tqdm(enumerate(lowerCamelCase ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCAmelCase = tokenizer( example.text_a , example.text_b , add_special_tokens=lowerCamelCase , max_length=lowerCamelCase , padding='max_length' , truncation=lowerCamelCase , return_overflowing_tokens=lowerCamelCase , ) lowerCAmelCase = label_map[example.label] if example.label in label_map else 0 lowerCAmelCase = int(example.pairID ) features.append(InputFeatures(**lowerCamelCase , label=lowerCamelCase , pairID=lowerCamelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(f'''guid: {example}''' ) logger.info(f'''features: {features[i]}''' ) return features __snake_case ={ """hans""": 3, } __snake_case ={ """hans""": HansProcessor, }
4
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''bert_for_seq_generation''': ( '''https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model''' ), } } lowerCAmelCase__ = {'''bert_for_seq_generation''': 512} class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : List[int] = [] SCREAMING_SNAKE_CASE : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[Any] ,lowercase__ : Tuple ,lowercase__ : Tuple="<s>" ,lowercase__ : Union[str, Any]="</s>" ,lowercase__ : str="<unk>" ,lowercase__ : Tuple="<pad>" ,lowercase__ : Union[str, Any]="<::::>" ,lowercase__ : Optional[Dict[str, Any]] = None ,**lowercase__ : Any ,): __lowercase = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,pad_token=lowercase__ ,sep_token=lowercase__ ,sp_model_kwargs=self.sp_model_kwargs ,**lowercase__ ,) __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.sp_model.get_piece_size() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : Optional[int] ,lowercase__ : Optional[Any] ): __lowercase = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs''' ): __lowercase = {} __lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : str ): return self.sp_model.encode(lowercase__ ,out_type=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Union[str, Any] ): return self.sp_model.piece_to_id(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.IdToPiece(lowercase__ ) return token def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : int ): __lowercase = [] __lowercase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowercase__ ) + token __lowercase = [] else: current_sub_tokens.append(lowercase__ ) out_string += self.sp_model.decode(lowercase__ ) return out_string.strip() def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,lowercase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowercase__ ,'''wb''' ) as fi: __lowercase = self.sp_model.serialized_model_proto() fi.write(lowercase__ ) return (out_vocab_file,)
104
0
import argparse SCREAMING_SNAKE_CASE__ = """docs/source/_static/js/custom.js""" def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Tuple ): '''simple docstring''' with open(__lowerCamelCase , encoding="utf-8" , newline="\n" ) as f: lowercase_ = f.readlines() lowercase_ = 0 # First let's put the right version while not lines[index].startswith("const stableVersion =" ): index += 1 lowercase_ = F'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith("const versionMapping = {" ): index += 1 # We go until the end while not lines[index].startswith("}" ): index += 1 # We add the new version at the end lines[index - 1] += F' "v{version}": "v{version}",\n' with open(__lowerCamelCase , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(__lowerCamelCase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") SCREAMING_SNAKE_CASE__ = parser.parse_args() update_custom_js(args.version)
359
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE__ = {"""configuration_mra""": ["""MRA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MraConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ """MRA_PRETRAINED_MODEL_ARCHIVE_LIST""", """MraForMaskedLM""", """MraForMultipleChoice""", """MraForQuestionAnswering""", """MraForSequenceClassification""", """MraForTokenClassification""", """MraLayer""", """MraModel""", """MraPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
297
0
"""simple docstring""" import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def lowercase ( a__ : int = 3 ) -> qiskit.result.counts.Counts: if isinstance(a__ , a__ ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(a__ ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) _UpperCamelCase = QuantumRegister(a__ , '''qr''' ) _UpperCamelCase = ClassicalRegister(a__ , '''cr''' ) _UpperCamelCase = QuantumCircuit(a__ , a__ ) _UpperCamelCase = number_of_qubits for i in range(a__ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(a__ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , a__ , a__ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(a__ , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(a__ , a__ ) # simulate with 10000 shots _UpperCamelCase = Aer.get_backend('''qasm_simulator''' ) _UpperCamelCase = execute(a__ , a__ , shots=10000 ) return job.result().get_counts(a__ ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
256
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"""vocab_file""": """spiece.model"""} UpperCAmelCase = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } UpperCAmelCase = {"""bert_for_seq_generation""": 512} class UpperCAmelCase_ ( _lowercase): snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = [] snake_case__ = ['''input_ids''', '''attention_mask'''] def __init__( self : Any , __UpperCamelCase : int , __UpperCamelCase : Optional[int]="<s>" , __UpperCamelCase : Optional[Any]="</s>" , __UpperCamelCase : Optional[Any]="<unk>" , __UpperCamelCase : Tuple="<pad>" , __UpperCamelCase : int="<::::>" , __UpperCamelCase : Optional[Dict[str, Any]] = None , **__UpperCamelCase : Any , ) -> None: _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , pad_token=__UpperCamelCase , sep_token=__UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCamelCase , ) _UpperCamelCase = vocab_file _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCamelCase ) @property def _UpperCamelCase ( self : Optional[int] ) -> Tuple: return self.sp_model.get_piece_size() def _UpperCamelCase ( self : int ) -> Optional[int]: _UpperCamelCase = {self.convert_ids_to_tokens(__UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ) -> Union[str, Any]: _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : str , __UpperCamelCase : Any ) -> Tuple: _UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : Union[str, Any] , __UpperCamelCase : str ) -> List[str]: return self.sp_model.encode(__UpperCamelCase , out_type=__UpperCamelCase ) def _UpperCamelCase ( self : Tuple , __UpperCamelCase : Any ) -> Optional[int]: return self.sp_model.piece_to_id(__UpperCamelCase ) def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : Optional[int] ) -> Optional[Any]: _UpperCamelCase = self.sp_model.IdToPiece(__UpperCamelCase ) return token def _UpperCamelCase ( self : str , __UpperCamelCase : Dict ) -> Optional[Any]: _UpperCamelCase = [] _UpperCamelCase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__UpperCamelCase ) + token _UpperCamelCase = [] else: current_sub_tokens.append(__UpperCamelCase ) out_string += self.sp_model.decode(__UpperCamelCase ) return out_string.strip() def _UpperCamelCase ( self : Tuple , __UpperCamelCase : str , __UpperCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__UpperCamelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase = os.path.join( __UpperCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCamelCase , '''wb''' ) as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCamelCase ) return (out_vocab_file,)
256
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase_ ={ """configuration_graphormer""": ["""GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GraphormerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ =[ """GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """GraphormerForGraphClassification""", """GraphormerModel""", """GraphormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys UpperCamelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
128
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class _a ( _lowerCAmelCase , _lowerCAmelCase ): @register_to_config def __init__( self : Any, lowerCAmelCase__ : int = 1_2_8, lowerCAmelCase__ : int = 2_5_6, lowerCAmelCase__ : float = 2_000.0, lowerCAmelCase__ : int = 7_6_8, lowerCAmelCase__ : int = 1_2, lowerCAmelCase__ : int = 1_2, lowerCAmelCase__ : int = 6_4, lowerCAmelCase__ : int = 2_0_4_8, lowerCAmelCase__ : float = 0.1, ) -> Any: '''simple docstring''' super().__init__() _UpperCamelCase : Any = nn.Sequential( nn.Linear(lowerCAmelCase__, d_model * 4, bias=lowerCAmelCase__ ), nn.SiLU(), nn.Linear(d_model * 4, d_model * 4, bias=lowerCAmelCase__ ), nn.SiLU(), ) _UpperCamelCase : List[Any] = nn.Embedding(lowerCAmelCase__, lowerCAmelCase__ ) _UpperCamelCase : List[Any] = False _UpperCamelCase : Optional[Any] = nn.Linear(lowerCAmelCase__, lowerCAmelCase__, bias=lowerCAmelCase__ ) _UpperCamelCase : Any = nn.Dropout(p=lowerCAmelCase__ ) _UpperCamelCase : List[str] = nn.ModuleList() for lyr_num in range(lowerCAmelCase__ ): # FiLM conditional T5 decoder _UpperCamelCase : Any = DecoderLayer(d_model=lowerCAmelCase__, d_kv=lowerCAmelCase__, num_heads=lowerCAmelCase__, d_ff=lowerCAmelCase__, dropout_rate=lowerCAmelCase__ ) self.decoders.append(lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = TaLayerNorm(lowerCAmelCase__ ) _UpperCamelCase : Dict = nn.Dropout(p=lowerCAmelCase__ ) _UpperCamelCase : Dict = nn.Linear(lowerCAmelCase__, lowerCAmelCase__, bias=lowerCAmelCase__ ) def snake_case ( self : Optional[Any], lowerCAmelCase__ : Union[str, Any], lowerCAmelCase__ : Optional[int] ) -> List[str]: '''simple docstring''' _UpperCamelCase : List[Any] = torch.mul(query_input.unsqueeze(-1 ), key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def snake_case ( self : str, lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[Any] ) -> Dict: '''simple docstring''' _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Union[str, Any] = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. _UpperCamelCase : Any = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time, embedding_dim=self.config.d_model, max_period=self.config.max_decoder_noise_time, ).to(dtype=self.dtype ) _UpperCamelCase : Union[str, Any] = self.conditioning_emb(lowerCAmelCase__ ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) _UpperCamelCase : int = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. _UpperCamelCase : int = torch.broadcast_to( torch.arange(lowerCAmelCase__, device=decoder_input_tokens.device ), (batch, seq_length), ) _UpperCamelCase : Dict = self.position_encoding(lowerCAmelCase__ ) _UpperCamelCase : List[str] = self.continuous_inputs_projection(lowerCAmelCase__ ) inputs += position_encodings _UpperCamelCase : Dict = self.dropout(lowerCAmelCase__ ) # decoder: No padding present. _UpperCamelCase : Tuple = torch.ones( decoder_input_tokens.shape[:2], device=decoder_input_tokens.device, dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. _UpperCamelCase : Tuple = [(x, self.encoder_decoder_mask(lowerCAmelCase__, lowerCAmelCase__ )) for x, y in encodings_and_masks] # cross attend style: concat encodings _UpperCamelCase : int = torch.cat([x[0] for x in encodings_and_encdec_masks], dim=1 ) _UpperCamelCase : Dict = torch.cat([x[1] for x in encodings_and_encdec_masks], dim=-1 ) for lyr in self.decoders: _UpperCamelCase : List[Any] = lyr( lowerCAmelCase__, conditioning_emb=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, encoder_attention_mask=lowerCAmelCase__, )[0] _UpperCamelCase : Any = self.decoder_norm(lowerCAmelCase__ ) _UpperCamelCase : Tuple = self.post_dropout(lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = self.spec_out(lowerCAmelCase__ ) return spec_out class _a ( nn.Module ): def __init__( self : Union[str, Any], lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : List[str], lowerCAmelCase__ : List[str], lowerCAmelCase__ : str, lowerCAmelCase__ : Union[str, Any]=1e-6 ) -> Optional[int]: '''simple docstring''' super().__init__() _UpperCamelCase : Optional[int] = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=lowerCAmelCase__, d_kv=lowerCAmelCase__, num_heads=lowerCAmelCase__, dropout_rate=lowerCAmelCase__ ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=lowerCAmelCase__, d_kv=lowerCAmelCase__, num_heads=lowerCAmelCase__, dropout_rate=lowerCAmelCase__, layer_norm_epsilon=lowerCAmelCase__, ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=lowerCAmelCase__, d_ff=lowerCAmelCase__, dropout_rate=lowerCAmelCase__, layer_norm_epsilon=lowerCAmelCase__ ) ) def snake_case ( self : Union[str, Any], lowerCAmelCase__ : str, lowerCAmelCase__ : List[Any]=None, lowerCAmelCase__ : List[str]=None, lowerCAmelCase__ : List[Any]=None, lowerCAmelCase__ : Optional[Any]=None, lowerCAmelCase__ : Any=None, ) -> List[Any]: '''simple docstring''' _UpperCamelCase : List[str] = self.layer[0]( lowerCAmelCase__, conditioning_emb=lowerCAmelCase__, attention_mask=lowerCAmelCase__, ) if encoder_hidden_states is not None: _UpperCamelCase : Any = torch.where(encoder_attention_mask > 0, 0, -1e1_0 ).to( encoder_hidden_states.dtype ) _UpperCamelCase : int = self.layer[1]( lowerCAmelCase__, key_value_states=lowerCAmelCase__, attention_mask=lowerCAmelCase__, ) # Apply Film Conditional Feed Forward layer _UpperCamelCase : Optional[int] = self.layer[-1](lowerCAmelCase__, lowerCAmelCase__ ) return (hidden_states,) class _a ( nn.Module ): def __init__( self : Tuple, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Tuple, lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : Optional[int] ) -> Optional[Any]: '''simple docstring''' super().__init__() _UpperCamelCase : int = TaLayerNorm(lowerCAmelCase__ ) _UpperCamelCase : List[Any] = TaFiLMLayer(in_features=d_model * 4, out_features=lowerCAmelCase__ ) _UpperCamelCase : Any = Attention(query_dim=lowerCAmelCase__, heads=lowerCAmelCase__, dim_head=lowerCAmelCase__, out_bias=lowerCAmelCase__, scale_qk=lowerCAmelCase__ ) _UpperCamelCase : List[str] = nn.Dropout(lowerCAmelCase__ ) def snake_case ( self : Optional[int], lowerCAmelCase__ : Tuple, lowerCAmelCase__ : str=None, lowerCAmelCase__ : Union[str, Any]=None, ) -> List[str]: '''simple docstring''' _UpperCamelCase : Any = self.layer_norm(lowerCAmelCase__ ) if conditioning_emb is not None: _UpperCamelCase : str = self.FiLMLayer(lowerCAmelCase__, lowerCAmelCase__ ) # Self-attention block _UpperCamelCase : Tuple = self.attention(lowerCAmelCase__ ) _UpperCamelCase : Dict = hidden_states + self.dropout(lowerCAmelCase__ ) return hidden_states class _a ( nn.Module ): def __init__( self : Tuple, lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : List[str], lowerCAmelCase__ : str, lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Tuple ) -> Tuple: '''simple docstring''' super().__init__() _UpperCamelCase : Any = Attention(query_dim=lowerCAmelCase__, heads=lowerCAmelCase__, dim_head=lowerCAmelCase__, out_bias=lowerCAmelCase__, scale_qk=lowerCAmelCase__ ) _UpperCamelCase : Tuple = TaLayerNorm(lowerCAmelCase__, eps=lowerCAmelCase__ ) _UpperCamelCase : int = nn.Dropout(lowerCAmelCase__ ) def snake_case ( self : Optional[Any], lowerCAmelCase__ : List[Any], lowerCAmelCase__ : str=None, lowerCAmelCase__ : Union[str, Any]=None, ) -> List[Any]: '''simple docstring''' _UpperCamelCase : Optional[int] = self.layer_norm(lowerCAmelCase__ ) _UpperCamelCase : Tuple = self.attention( lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, attention_mask=attention_mask.squeeze(1 ), ) _UpperCamelCase : str = hidden_states + self.dropout(lowerCAmelCase__ ) return layer_output class _a ( nn.Module ): def __init__( self : Tuple, lowerCAmelCase__ : int, lowerCAmelCase__ : Tuple, lowerCAmelCase__ : Any, lowerCAmelCase__ : str ) -> Tuple: '''simple docstring''' super().__init__() _UpperCamelCase : Any = TaDenseGatedActDense(d_model=lowerCAmelCase__, d_ff=lowerCAmelCase__, dropout_rate=lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = TaFiLMLayer(in_features=d_model * 4, out_features=lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = TaLayerNorm(lowerCAmelCase__, eps=lowerCAmelCase__ ) _UpperCamelCase : List[Any] = nn.Dropout(lowerCAmelCase__ ) def snake_case ( self : Tuple, lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : List[str]=None ) -> Tuple: '''simple docstring''' _UpperCamelCase : Any = self.layer_norm(lowerCAmelCase__ ) if conditioning_emb is not None: _UpperCamelCase : Dict = self.film(lowerCAmelCase__, lowerCAmelCase__ ) _UpperCamelCase : str = self.DenseReluDense(lowerCAmelCase__ ) _UpperCamelCase : str = hidden_states + self.dropout(lowerCAmelCase__ ) return hidden_states class _a ( nn.Module ): def __init__( self : str, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Dict ) -> Tuple: '''simple docstring''' super().__init__() _UpperCamelCase : List[str] = nn.Linear(lowerCAmelCase__, lowerCAmelCase__, bias=lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = nn.Linear(lowerCAmelCase__, lowerCAmelCase__, bias=lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = nn.Linear(lowerCAmelCase__, lowerCAmelCase__, bias=lowerCAmelCase__ ) _UpperCamelCase : List[str] = nn.Dropout(lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = NewGELUActivation() def snake_case ( self : str, lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.act(self.wi_a(lowerCAmelCase__ ) ) _UpperCamelCase : Dict = self.wi_a(lowerCAmelCase__ ) _UpperCamelCase : Any = hidden_gelu * hidden_linear _UpperCamelCase : Any = self.dropout(lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = self.wo(lowerCAmelCase__ ) return hidden_states class _a ( nn.Module ): def __init__( self : Union[str, Any], lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : Optional[Any]=1e-6 ) -> Optional[Any]: '''simple docstring''' super().__init__() _UpperCamelCase : Optional[Any] = nn.Parameter(torch.ones(lowerCAmelCase__ ) ) _UpperCamelCase : Tuple = eps def snake_case ( self : Dict, lowerCAmelCase__ : Tuple ) -> Any: '''simple docstring''' _UpperCamelCase : Optional[Any] = hidden_states.to(torch.floataa ).pow(2 ).mean(-1, keepdim=lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: _UpperCamelCase : int = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class _a ( nn.Module ): def snake_case ( self : Optional[Any], lowerCAmelCase__ : torch.Tensor ) -> torch.Tensor: '''simple docstring''' return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.044_715 * torch.pow(lowerCAmelCase__, 3.0 )) )) class _a ( nn.Module ): def __init__( self : List[str], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : int ) -> List[Any]: '''simple docstring''' super().__init__() _UpperCamelCase : Tuple = nn.Linear(lowerCAmelCase__, out_features * 2, bias=lowerCAmelCase__ ) def snake_case ( self : int, lowerCAmelCase__ : List[str], lowerCAmelCase__ : Any ) -> int: '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.scale_bias(lowerCAmelCase__ ) _UpperCamelCase , _UpperCamelCase : Union[str, Any] = torch.chunk(lowerCAmelCase__, 2, -1 ) _UpperCamelCase : Optional[int] = x * (1 + scale) + shift return x
128
1
"""simple docstring""" import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() UpperCamelCase_ = 2 class snake_case : def __init__( self , *, # begin keyword-only arguments __UpperCAmelCase="<s>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="<unk>" , __UpperCAmelCase=None , ) ->Tuple: a_ , a_ , a_ , a_ = bos, unk, pad, eos a_ = [] a_ = [] a_ = {} a_ = self.add_symbol(__UpperCAmelCase) a_ = self.add_symbol(__UpperCAmelCase) a_ = self.add_symbol(__UpperCAmelCase) a_ = self.add_symbol(__UpperCAmelCase) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(__UpperCAmelCase) a_ = len(self.symbols) def __eq__( self , __UpperCAmelCase) ->Dict: return self.indices == other.indices def __getitem__( self , __UpperCAmelCase) ->Optional[Any]: if idx < len(self.symbols): return self.symbols[idx] return self.unk_word def __len__( self) ->Any: return len(self.symbols) def __contains__( self , __UpperCAmelCase) ->Dict: return sym in self.indices @classmethod def UpperCAmelCase__ ( cls , __UpperCAmelCase) ->List[Any]: a_ = cls() d.add_from_file(__UpperCAmelCase) return d def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase=1 , __UpperCAmelCase=False) ->List[Any]: if word in self.indices and not overwrite: a_ = self.indices[word] a_ = self.count[idx] + n return idx else: a_ = len(self.symbols) a_ = idx self.symbols.append(__UpperCAmelCase) self.count.append(__UpperCAmelCase) return idx def UpperCAmelCase__ ( self , __UpperCAmelCase) ->Tuple: return 0 def UpperCAmelCase__ ( self , __UpperCAmelCase) ->List[str]: if isinstance(__UpperCAmelCase , __UpperCAmelCase): try: with open(__UpperCAmelCase , "r" , encoding="utf-8") as fd: self.add_from_file(__UpperCAmelCase) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception("Incorrect encoding detected in {}, please rebuild the dataset".format(__UpperCAmelCase)) return a_ = f.readlines() a_ = self._load_meta(__UpperCAmelCase) for line in lines[indices_start_line:]: try: a_ , a_ = line.rstrip().rsplit(" " , 1) if field == "#fairseq:overwrite": a_ = True a_ , a_ = line.rsplit(" " , 1) else: a_ = False a_ = int(__UpperCAmelCase) a_ = line if word in self and not overwrite: raise RuntimeError( "Duplicate word found when loading Dictionary: '{}'. " "Duplicate words can overwrite earlier ones by adding the " "#fairseq:overwrite flag at the end of the corresponding row " "in the dictionary file. If using the Camembert model, please " "download an updated copy of the model file.".format(__UpperCAmelCase)) self.add_symbol(__UpperCAmelCase , n=__UpperCAmelCase , overwrite=__UpperCAmelCase) except ValueError: raise ValueError("Incorrect dictionary format, expected '<token> <cnt> [flags]'") def UpperCamelCase ( UpperCAmelCase ) ->Union[str, Any]: """simple docstring""" a_ = dict((re.sub(r"@@$" , "" , UpperCAmelCase ), v) if k.endswith("@@" ) else (re.sub(r"$" , "</w>" , UpperCAmelCase ), v) for k, v in d.items() ) a_ = "<s> <pad> </s> <unk>".split() # restore the special tokens for k in keep_keys: del da[F'''{k}</w>'''] a_ = d[k] # restore return da def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ) ->Any: """simple docstring""" if not os.path.exists(UpperCAmelCase ): raise ValueError(F'''path {biogpt_checkpoint_path} does not exist!''' ) os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) print(F'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models a_ = os.path.join(UpperCAmelCase , "checkpoint.pt" ) if not os.path.isfile(UpperCAmelCase ): raise ValueError(F'''path to the file {checkpoint_file} does not exist!''' ) a_ = torch.load(UpperCAmelCase , map_location="cpu" ) a_ = chkpt["cfg"]["model"] # dicts a_ = os.path.join(UpperCAmelCase , "dict.txt" ) if not os.path.isfile(UpperCAmelCase ): raise ValueError(F'''path to the file {dict_file} does not exist!''' ) a_ = Dictionary.load(UpperCAmelCase ) a_ = rewrite_dict_keys(src_dict.indices ) a_ = len(UpperCAmelCase ) a_ = os.path.join(UpperCAmelCase , VOCAB_FILES_NAMES["vocab_file"] ) print(F'''Generating {src_vocab_file} of {src_vocab_size} records''' ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(UpperCAmelCase , ensure_ascii=UpperCAmelCase , indent=UpperCAmelCase ) ) # merges_file (bpecodes) a_ = os.path.join(UpperCAmelCase , "bpecodes" ) if not os.path.isfile(UpperCAmelCase ): raise ValueError(F'''path to the file {bpecodes_file} does not exist!''' ) a_ = os.path.join(UpperCAmelCase , VOCAB_FILES_NAMES["merges_file"] ) shutil.copyfile(UpperCAmelCase , UpperCAmelCase ) # model config a_ = os.path.join(UpperCAmelCase , "config.json" ) a_ = { "activation_dropout": args["activation_dropout"], "architectures": ["BioGptForCausalLM"], "attention_probs_dropout_prob": args["attention_dropout"], "bos_token_id": 0, "eos_token_id": 2, "hidden_act": args["activation_fn"], "hidden_dropout_prob": args["dropout"], "hidden_size": args["decoder_embed_dim"], "initializer_range": 0.02, "intermediate_size": args["decoder_ffn_embed_dim"], "layer_norm_eps": 1E-12, "layerdrop": args["decoder_layerdrop"], "max_position_embeddings": args["max_target_positions"], "model_type": "biogpt", "num_attention_heads": args["decoder_attention_heads"], "num_hidden_layers": args["decoder_layers"], "pad_token_id": 1, "scale_embedding": not args["no_scale_embedding"], "tie_word_embeddings": args["share_decoder_input_output_embed"], "vocab_size": src_vocab_size, } # good hparam defaults to start with print(F'''Generating {biogpt_model_config_file}''' ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(UpperCAmelCase , ensure_ascii=UpperCAmelCase , indent=UpperCAmelCase ) ) # tokenizer config a_ = os.path.join(UpperCAmelCase , UpperCAmelCase ) a_ = { "bos_token": "<s>", "eos_token": "</s>", "model_max_length": 1_024, "pad_token": "<pad>", "special_tokens_map_file": None, "tokenizer_class": "BioGptTokenizer", "unk_token": "<unk>", } print(F'''Generating {biogpt_tokenizer_config_file}''' ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(UpperCAmelCase , ensure_ascii=UpperCAmelCase , indent=UpperCAmelCase ) ) # model a_ = chkpt["model"] # remove unneeded keys a_ = [ "decoder.version", ] for k in ignore_keys: model_state_dict.pop(UpperCAmelCase , UpperCAmelCase ) a_ = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith("output_projection.weight" ): a_ = model_state_dict.pop(UpperCAmelCase ) else: a_ = model_state_dict.pop(UpperCAmelCase ) a_ = BioGptConfig.from_pretrained(UpperCAmelCase ) a_ = BioGptForCausalLM(UpperCAmelCase ) # check that it loads ok model_new.load_state_dict(UpperCAmelCase ) # save a_ = os.path.join(UpperCAmelCase , UpperCAmelCase ) print(F'''Generating {pytorch_weights_dump_path}''' ) torch.save(UpperCAmelCase , UpperCAmelCase ) print("Conversion is done!" ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--biogpt_checkpoint_path', default=None, type=str, required=True, help=( 'Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,' ' bpecodes, etc.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) UpperCamelCase_ = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
243
"""simple docstring""" def UpperCamelCase ( UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) ->bool: """simple docstring""" a_ = set() # Replace all the whitespace in our sentence a_ = input_str.replace(" " , "" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(UpperCAmelCase ) == 26 def UpperCamelCase ( UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) ->bool: """simple docstring""" a_ = [False] * 26 for char in input_str: if char.islower(): a_ = True elif char.isupper(): a_ = True return all(UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) ->bool: """simple docstring""" return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def UpperCamelCase ( ) ->None: """simple docstring""" from timeit import timeit a_ = "from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest" print(timeit("is_pangram()" , setup=UpperCAmelCase ) ) print(timeit("is_pangram_faster()" , setup=UpperCAmelCase ) ) print(timeit("is_pangram_fastest()" , setup=UpperCAmelCase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
243
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json" ), } class snake_case ( _UpperCamelCase): __UpperCamelCase = 'dpr' def __init__( self : int , a__ : Optional[int]=3_05_22 , a__ : Union[str, Any]=7_68 , a__ : Optional[Any]=12 , a__ : str=12 , a__ : Union[str, Any]=30_72 , a__ : List[str]="gelu" , a__ : List[Any]=0.1 , a__ : Dict=0.1 , a__ : Optional[int]=5_12 , a__ : str=2 , a__ : Dict=0.0_2 , a__ : Any=1E-1_2 , a__ : List[Any]=0 , a__ : Dict="absolute" , a__ : Dict = 0 , **a__ : str , ) -> Tuple: '''simple docstring''' super().__init__(pad_token_id=__a , **__a ) _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = hidden_act _A = intermediate_size _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = initializer_range _A = layer_norm_eps _A = projection_dim _A = position_embedding_type
362
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase): __UpperCamelCase = StableDiffusionInstructPixaPixPipeline __UpperCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width', 'cross_attention_kwargs'} __UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS __UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def a_ ( self : Optional[int] ) -> str: '''simple docstring''' torch.manual_seed(0 ) _A = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) _A = PNDMScheduler(skip_prk_steps=a__ ) torch.manual_seed(0 ) _A = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) _A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _A = CLIPTextModel(a__ ) _A = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _A = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def a_ ( self : Optional[Any] , a__ : Dict , a__ : Tuple=0 ) -> Union[str, Any]: '''simple docstring''' _A = floats_tensor((1, 3, 32, 32) , rng=random.Random(a__ ) ).to(a__ ) _A = image.cpu().permute(0 , 2 , 3 , 1 )[0] _A = Image.fromarray(np.uinta(a__ ) ).convert("RGB" ) if str(a__ ).startswith("mps" ): _A = torch.manual_seed(a__ ) else: _A = torch.Generator(device=a__ ).manual_seed(a__ ) _A = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "image_guidance_scale": 1, "output_type": "numpy", } return inputs def a_ ( self : Dict ) -> str: '''simple docstring''' _A = "cpu" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = StableDiffusionInstructPixaPixPipeline(**a__ ) _A = sd_pipe.to(a__ ) sd_pipe.set_progress_bar_config(disable=a__ ) _A = self.get_dummy_inputs(a__ ) _A = sd_pipe(**a__ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _A = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a_ ( self : str ) -> Optional[int]: '''simple docstring''' _A = "cpu" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = StableDiffusionInstructPixaPixPipeline(**a__ ) _A = sd_pipe.to(a__ ) sd_pipe.set_progress_bar_config(disable=a__ ) _A = self.get_dummy_inputs(a__ ) _A = "french fries" _A = sd_pipe(**a__ , negative_prompt=a__ ) _A = output.images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _A = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a_ ( self : Optional[int] ) -> int: '''simple docstring''' _A = "cpu" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = StableDiffusionInstructPixaPixPipeline(**a__ ) _A = sd_pipe.to(a__ ) sd_pipe.set_progress_bar_config(disable=a__ ) _A = self.get_dummy_inputs(a__ ) _A = [inputs["prompt"]] * 2 _A = np.array(inputs["image"] ).astype(np.floataa ) / 2_5_5.0 _A = torch.from_numpy(a__ ).unsqueeze(0 ).to(a__ ) _A = image / 2 + 0.5 _A = image.permute(0 , 3 , 1 , 2 ) _A = image.repeat(2 , 1 , 1 , 1 ) _A = sd_pipe(**a__ ).images _A = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) _A = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' _A = "cpu" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" ) _A = StableDiffusionInstructPixaPixPipeline(**a__ ) _A = sd_pipe.to(a__ ) sd_pipe.set_progress_bar_config(disable=a__ ) _A = self.get_dummy_inputs(a__ ) _A = sd_pipe(**a__ ).images _A = image[0, -3:, -3:, -1] _A = [round(a__ , 4 ) for x in image_slice.flatten().tolist()] print(",".join([str(a__ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) _A = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a_ ( self : List[str] ) -> int: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def a_ ( self : str ) -> Any: '''simple docstring''' _A = self.get_dummy_components() _A = StableDiffusionInstructPixaPixPipeline(**a__ ) _A = VaeImageProcessor(do_resize=a__ , do_normalize=a__ ) _A = pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) _A = pipe(**self.get_dummy_inputs_by_type(a__ , input_image_type="pt" ) )[0] _A = components["vae"] _A = self.get_dummy_inputs_by_type(a__ , input_image_type="pt" ) for image_param in self.image_latents_params: if image_param in inputs.keys(): _A = vae.encode(inputs[image_param] ).latent_dist.mode() _A = pipe(**a__ )[0] _A = np.abs(out - out_latents_inputs ).max() self.assertLess(a__ , 1E-4 , "passing latents as image input generate different result from passing image" ) @slow @require_torch_gpu class snake_case ( unittest.TestCase): def a_ ( self : List[str] ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self : Optional[Any] , a__ : str=0 ) -> List[Any]: '''simple docstring''' _A = torch.manual_seed(a__ ) _A = load_image( "https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg" ) _A = { "prompt": "turn him into a cyborg", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "image_guidance_scale": 1.0, "output_type": "numpy", } return inputs def a_ ( self : List[Any] ) -> Any: '''simple docstring''' _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=a__ ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() _A = self.get_inputs() _A = pipe(**a__ ).images _A = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) _A = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def a_ ( self : List[Any] ) -> Any: '''simple docstring''' _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=a__ ) _A = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() _A = self.get_inputs() _A = pipe(**a__ ).images _A = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) _A = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def a_ ( self : Union[str, Any] ) -> int: '''simple docstring''' _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=a__ ) _A = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() _A = self.get_inputs() _A = pipe(**a__ ).images _A = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) _A = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def a_ ( self : Optional[int] ) -> Dict: '''simple docstring''' _A = 0 def callback_fn(a__ : int , a__ : int , a__ : torch.FloatTensor ) -> None: _A = True nonlocal number_of_steps number_of_steps += 1 if step == 1: _A = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) _A = latents[0, -3:, -3:, -1] _A = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: _A = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) _A = latents[0, -3:, -3:, -1] _A = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 _A = False _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=a__ , torch_dtype=torch.floataa ) _A = pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() _A = self.get_inputs() pipe(**a__ , callback=a__ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def a_ ( self : List[Any] ) -> Any: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=a__ , torch_dtype=torch.floataa ) _A = pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _A = self.get_inputs() _A = pipe(**a__ ) _A = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def a_ ( self : List[Any] ) -> List[Any]: '''simple docstring''' _A = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 _A = inputs["image"].resize((5_04, 5_04) ) _A = "timbrooks/instruct-pix2pix" _A = StableDiffusionInstructPixaPixPipeline.from_pretrained( a__ , safety_checker=a__ , ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() _A = pipe(**a__ ) _A = output.images[0] _A = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 5_04, 3) _A = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
163
0
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class __lowerCamelCase (unittest.TestCase ): def __init__( self: Any,A_: Union[str, Any],A_: Union[str, Any]=13,A_: List[Any]=7,A_: Tuple=True,A_: Optional[Any]=True,A_: Dict=True,A_: Union[str, Any]=True,A_: List[Any]=99,A_: str=32,A_: List[str]=5,A_: Optional[int]=4,A_: List[Any]=37,A_: List[str]="gelu",A_: Any=0.1,A_: List[Any]=0.1,A_: Dict=512,A_: Optional[int]=16,A_: List[str]=2,A_: Optional[Any]=0.0_2,A_: Optional[Any]=4,): '''simple docstring''' __UpperCamelCase = parent __UpperCamelCase = batch_size __UpperCamelCase = seq_length __UpperCamelCase = is_training __UpperCamelCase = use_attention_mask __UpperCamelCase = use_token_type_ids __UpperCamelCase = use_labels __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = type_sequence_label_size __UpperCamelCase = initializer_range __UpperCamelCase = num_choices def snake_case_ ( self: List[Any] ): '''simple docstring''' __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length],self.vocab_size ) __UpperCamelCase = None if self.use_attention_mask: __UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase = None if self.use_token_type_ids: __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length],self.type_vocab_size ) __UpperCamelCase = RobertaConfig( vocab_size=self.vocab_size,hidden_size=self.hidden_size,num_hidden_layers=self.num_hidden_layers,num_attention_heads=self.num_attention_heads,intermediate_size=self.intermediate_size,hidden_act=self.hidden_act,hidden_dropout_prob=self.hidden_dropout_prob,attention_probs_dropout_prob=self.attention_probs_dropout_prob,max_position_embeddings=self.max_position_embeddings,type_vocab_size=self.type_vocab_size,is_decoder=A_,initializer_range=self.initializer_range,) return config, input_ids, token_type_ids, attention_mask def snake_case_ ( self: str ): '''simple docstring''' __UpperCamelCase = self.prepare_config_and_inputs() __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase = config_and_inputs __UpperCamelCase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def snake_case_ ( self: Any ): '''simple docstring''' __UpperCamelCase = self.prepare_config_and_inputs() __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase = config_and_inputs __UpperCamelCase = True __UpperCamelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length],vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class __lowerCamelCase (_a , unittest.TestCase ): _lowercase = True _lowercase = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def snake_case_ ( self: Any ): '''simple docstring''' __UpperCamelCase = FlaxRobertaModelTester(self ) @slow def snake_case_ ( self: Dict ): '''simple docstring''' for model_class_name in self.all_model_classes: __UpperCamelCase = model_class_name.from_pretrained('roberta-base',from_pt=A_ ) __UpperCamelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(A_ )
310
def _A ( _lowercase ) -> list: """simple docstring""" def merge(_lowercase , _lowercase ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(_lowercase ) <= 1: return collection __UpperCamelCase = len(_lowercase ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() __snake_case = input('''Enter numbers separated by a comma:\n''').strip() __snake_case = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
310
1
"""simple docstring""" from ... import PretrainedConfig lowerCAmelCase__ = { '''sijunhe/nezha-cn-base''': '''https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json''', } class _lowerCamelCase ( _lowercase ): UpperCAmelCase_ = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP UpperCAmelCase_ = "nezha" def __init__(self , __a=2_11_28 , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=64 , __a=2 , __a=0.02 , __a=1e-1_2 , __a=0.1 , __a=0 , __a=2 , __a=3 , __a=True , **__a , ) -> str: super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = hidden_act UpperCamelCase = intermediate_size UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = max_relative_position UpperCamelCase = type_vocab_size UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = classifier_dropout UpperCamelCase = use_cache
244
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class _lowerCamelCase ( _lowercase ): UpperCAmelCase_ = "facebook/bart-large-mnli" UpperCAmelCase_ = ( "This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which " "should be the text to classify, and `labels`, which should be the list of labels to use for classification. " "It returns the most likely label in the list of provided `labels` for the input text." ) UpperCAmelCase_ = "text_classifier" UpperCAmelCase_ = AutoTokenizer UpperCAmelCase_ = AutoModelForSequenceClassification UpperCAmelCase_ = ["text", ["text"]] UpperCAmelCase_ = ["text"] def snake_case_ (self ) -> List[Any]: super().setup() UpperCamelCase = self.model.config UpperCamelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("entail" ): UpperCamelCase = int(__a ) if self.entailment_id == -1: raise ValueError("Could not determine the entailment ID from the model config, please pass it at init." ) def snake_case_ (self , __a , __a ) -> List[Any]: UpperCamelCase = labels return self.pre_processor( [text] * len(__a ) , [F"This example is {label}" for label in labels] , return_tensors="pt" , padding="max_length" , ) def snake_case_ (self , __a ) -> int: UpperCamelCase = outputs.logits UpperCamelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
244
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : List[Any] = logging.get_logger(__name__) lowerCamelCase : List[str] = { "studio-ousia/luke-base": "https://huggingface.co/studio-ousia/luke-base/resolve/main/config.json", "studio-ousia/luke-large": "https://huggingface.co/studio-ousia/luke-large/resolve/main/config.json", } class A__ ( A__ ): A__ = 'luke' def __init__( self : str , _a : Optional[int]=5_0267 , _a : Optional[int]=50_0000 , _a : int=768 , _a : str=256 , _a : int=12 , _a : Optional[int]=12 , _a : List[Any]=3072 , _a : Union[str, Any]="gelu" , _a : Optional[Any]=0.1 , _a : int=0.1 , _a : List[str]=512 , _a : Tuple=2 , _a : Union[str, Any]=0.02 , _a : Optional[int]=1e-12 , _a : Optional[Any]=True , _a : Any=None , _a : List[Any]=1 , _a : List[Any]=0 , _a : int=2 , **_a : str , ) -> Optional[int]: '''simple docstring''' super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) _SCREAMING_SNAKE_CASE =vocab_size _SCREAMING_SNAKE_CASE =entity_vocab_size _SCREAMING_SNAKE_CASE =hidden_size _SCREAMING_SNAKE_CASE =entity_emb_size _SCREAMING_SNAKE_CASE =num_hidden_layers _SCREAMING_SNAKE_CASE =num_attention_heads _SCREAMING_SNAKE_CASE =hidden_act _SCREAMING_SNAKE_CASE =intermediate_size _SCREAMING_SNAKE_CASE =hidden_dropout_prob _SCREAMING_SNAKE_CASE =attention_probs_dropout_prob _SCREAMING_SNAKE_CASE =max_position_embeddings _SCREAMING_SNAKE_CASE =type_vocab_size _SCREAMING_SNAKE_CASE =initializer_range _SCREAMING_SNAKE_CASE =layer_norm_eps _SCREAMING_SNAKE_CASE =use_entity_aware_attention _SCREAMING_SNAKE_CASE =classifier_dropout
47
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase : List[Any] = version.parse(importlib_metadata.version("nltk")) if NLTK_VERSION >= version.Version("3.6.4"): from nltk import word_tokenize lowerCamelCase : Any = "\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n" lowerCamelCase : Optional[Any] = "\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n" lowerCamelCase : Optional[Any] = "\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n 'meteor': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric('meteor')\n >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\n >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results[\"meteor\"], 4))\n 0.6944\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def A ( self : Tuple ) -> str: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def A ( self : Union[str, Any] , _a : Union[str, Any] ) -> Optional[int]: '''simple docstring''' import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def A ( self : int , _a : Tuple , _a : List[str] , _a : List[str]=0.9 , _a : Dict=3 , _a : Optional[int]=0.5 ) -> Optional[int]: '''simple docstring''' if NLTK_VERSION >= version.Version('3.6.5' ): _SCREAMING_SNAKE_CASE =[ meteor_score.single_meteor_score( word_tokenize(_a ) , word_tokenize(_a ) , alpha=_a , beta=_a , gamma=_a ) for ref, pred in zip(_a , _a ) ] else: _SCREAMING_SNAKE_CASE =[ meteor_score.single_meteor_score(_a , _a , alpha=_a , beta=_a , gamma=_a ) for ref, pred in zip(_a , _a ) ] return {"meteor": np.mean(_a )}
47
1
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch lowerCamelCase_ = random.Random() def __lowerCamelCase ( a_ : Optional[int] , a_ : Any=1.0 , a_ : Dict=None , a_ : List[str]=None ) -> Union[str, Any]: if rng is None: __SCREAMING_SNAKE_CASE :Tuple = global_rng __SCREAMING_SNAKE_CASE :str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def __init__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=7 ,SCREAMING_SNAKE_CASE__=4_00 ,SCREAMING_SNAKE_CASE__=20_00 ,SCREAMING_SNAKE_CASE__=1 ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=1_60_00 ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=80 ,SCREAMING_SNAKE_CASE__=16 ,SCREAMING_SNAKE_CASE__=64 ,SCREAMING_SNAKE_CASE__="hann_window" ,SCREAMING_SNAKE_CASE__=80 ,SCREAMING_SNAKE_CASE__=76_00 ,SCREAMING_SNAKE_CASE__=1E-10 ,SCREAMING_SNAKE_CASE__=True ,) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = parent __SCREAMING_SNAKE_CASE :str = batch_size __SCREAMING_SNAKE_CASE :List[str] = min_seq_length __SCREAMING_SNAKE_CASE :Optional[Any] = max_seq_length __SCREAMING_SNAKE_CASE :str = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __SCREAMING_SNAKE_CASE :Tuple = feature_size __SCREAMING_SNAKE_CASE :List[str] = padding_value __SCREAMING_SNAKE_CASE :List[str] = sampling_rate __SCREAMING_SNAKE_CASE :Union[str, Any] = do_normalize __SCREAMING_SNAKE_CASE :Optional[Any] = num_mel_bins __SCREAMING_SNAKE_CASE :int = hop_length __SCREAMING_SNAKE_CASE :Optional[Any] = win_length __SCREAMING_SNAKE_CASE :Tuple = win_function __SCREAMING_SNAKE_CASE :Any = fmin __SCREAMING_SNAKE_CASE :str = fmax __SCREAMING_SNAKE_CASE :str = mel_floor __SCREAMING_SNAKE_CASE :Any = return_attention_mask def _UpperCamelCase ( self ) -> int: """simple docstring""" return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=False ) -> List[str]: """simple docstring""" def _flatten(SCREAMING_SNAKE_CASE__ ): return list(itertools.chain(*lowercase_ ) ) if equal_length: __SCREAMING_SNAKE_CASE :List[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __SCREAMING_SNAKE_CASE :Optional[int] = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: __SCREAMING_SNAKE_CASE :Optional[Any] = [np.asarray(lowercase_ ) for x in speech_inputs] return speech_inputs def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=False ) -> Dict: """simple docstring""" if equal_length: __SCREAMING_SNAKE_CASE :Optional[Any] = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __SCREAMING_SNAKE_CASE :Union[str, Any] = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: __SCREAMING_SNAKE_CASE :Tuple = [np.asarray(lowercase_ ) for x in speech_inputs] return speech_inputs @require_torch class _SCREAMING_SNAKE_CASE( _UpperCAmelCase , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : int = SpeechTaFeatureExtractor def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = SpeechTaFeatureExtractionTester(self ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: """simple docstring""" self.assertTrue(np.all(np.mean(lowercase_ ,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowercase_ ,axis=0 ) - 1 ) < 1E-3 ) ) def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __SCREAMING_SNAKE_CASE :int = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :Any = [np.asarray(lowercase_ ) for speech_input in speech_inputs] # Test not batched input __SCREAMING_SNAKE_CASE :Optional[int] = feat_extract(speech_inputs[0] ,return_tensors='''np''' ).input_values __SCREAMING_SNAKE_CASE :Any = feat_extract(np_speech_inputs[0] ,return_tensors='''np''' ).input_values self.assertTrue(np.allclose(lowercase_ ,lowercase_ ,atol=1E-3 ) ) # Test batched __SCREAMING_SNAKE_CASE :Dict = feat_extract(lowercase_ ,return_tensors='''np''' ).input_values __SCREAMING_SNAKE_CASE :Optional[Any] = feat_extract(lowercase_ ,return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(lowercase_ ,lowercase_ ): self.assertTrue(np.allclose(lowercase_ ,lowercase_ ,atol=1E-3 ) ) def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE :int = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :List[Any] = ["""longest""", """max_length""", """do_not_pad"""] __SCREAMING_SNAKE_CASE :int = [None, 16_00, None] for max_length, padding in zip(lowercase_ ,lowercase_ ): __SCREAMING_SNAKE_CASE :Any = feat_extract(lowercase_ ,padding=lowercase_ ,max_length=lowercase_ ,return_tensors='''np''' ) __SCREAMING_SNAKE_CASE :Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self.assertTrue(input_values[0][8_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self.assertTrue(input_values[0][10_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE :Any = range(8_00 ,14_00 ,2_00 ) __SCREAMING_SNAKE_CASE :int = [floats_list((1, x) )[0] for x in lengths] __SCREAMING_SNAKE_CASE :int = ["""longest""", """max_length""", """do_not_pad"""] __SCREAMING_SNAKE_CASE :str = [None, 16_00, None] for max_length, padding in zip(lowercase_ ,lowercase_ ): __SCREAMING_SNAKE_CASE :Tuple = feat_extract(lowercase_ ,max_length=lowercase_ ,padding=lowercase_ ) __SCREAMING_SNAKE_CASE :Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def _UpperCamelCase ( self ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE :List[Any] = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :int = feat_extract( lowercase_ ,truncation=lowercase_ ,max_length=10_00 ,padding='''max_length''' ,return_tensors='''np''' ) __SCREAMING_SNAKE_CASE :int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE :Optional[Any] = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :str = feat_extract( lowercase_ ,truncation=lowercase_ ,max_length=10_00 ,padding='''longest''' ,return_tensors='''np''' ) __SCREAMING_SNAKE_CASE :List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00) ) __SCREAMING_SNAKE_CASE :List[str] = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :Optional[int] = feat_extract( lowercase_ ,truncation=lowercase_ ,max_length=20_00 ,padding='''longest''' ,return_tensors='''np''' ) __SCREAMING_SNAKE_CASE :int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00) ) def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __SCREAMING_SNAKE_CASE :Tuple = np.random.rand(1_00 ).astype(np.floataa ) __SCREAMING_SNAKE_CASE :Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __SCREAMING_SNAKE_CASE :List[Any] = feature_extractor.pad([{'''input_values''': inputs}] ,return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __SCREAMING_SNAKE_CASE :Any = feature_extractor.pad([{'''input_values''': inputs}] ,return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __SCREAMING_SNAKE_CASE :List[Any] = [floats_list((1, x) )[0] for x in range(8_00 ,14_00 ,2_00 )] __SCREAMING_SNAKE_CASE :Tuple = [np.asarray(lowercase_ ) for speech_input in speech_inputs] # Test feature size __SCREAMING_SNAKE_CASE :Dict = feature_extractor(audio_target=lowercase_ ,padding=lowercase_ ,return_tensors='''np''' ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input __SCREAMING_SNAKE_CASE :Optional[int] = feature_extractor(speech_inputs[0] ,return_tensors='''np''' ).input_values __SCREAMING_SNAKE_CASE :Any = feature_extractor(np_speech_inputs[0] ,return_tensors='''np''' ).input_values self.assertTrue(np.allclose(lowercase_ ,lowercase_ ,atol=1E-3 ) ) # Test batched __SCREAMING_SNAKE_CASE :Dict = feature_extractor(lowercase_ ,return_tensors='''np''' ).input_values __SCREAMING_SNAKE_CASE :int = feature_extractor(lowercase_ ,return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(lowercase_ ,lowercase_ ): self.assertTrue(np.allclose(lowercase_ ,lowercase_ ,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __SCREAMING_SNAKE_CASE :str = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __SCREAMING_SNAKE_CASE :str = np.asarray(lowercase_ ) __SCREAMING_SNAKE_CASE :Any = feature_extractor(lowercase_ ,return_tensors='''np''' ).input_values __SCREAMING_SNAKE_CASE :Dict = feature_extractor(lowercase_ ,return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(lowercase_ ,lowercase_ ): self.assertTrue(np.allclose(lowercase_ ,lowercase_ ,atol=1E-3 ) ) def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = self.feat_extract_tester.prepare_inputs_for_target() __SCREAMING_SNAKE_CASE :List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) __SCREAMING_SNAKE_CASE :List[str] = feat_extract.model_input_names[0] __SCREAMING_SNAKE_CASE :Dict = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowercase_ ) == len(lowercase_ ) for x, y in zip(lowercase_ ,processed_features[input_name] ) ) ) __SCREAMING_SNAKE_CASE :List[Any] = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowercase_ ) __SCREAMING_SNAKE_CASE :List[str] = BatchFeature({input_name: speech_inputs} ,tensor_type='''np''' ) __SCREAMING_SNAKE_CASE :List[Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: __SCREAMING_SNAKE_CASE :List[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowercase_ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) __SCREAMING_SNAKE_CASE :Union[str, Any] = feat_extract.model_input_names[0] __SCREAMING_SNAKE_CASE :List[Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='''pt''' ) __SCREAMING_SNAKE_CASE :Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: __SCREAMING_SNAKE_CASE :List[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = self.feature_extraction_class(**self.feat_extract_dict ) __SCREAMING_SNAKE_CASE :List[str] = self.feat_extract_tester.prepare_inputs_for_target() __SCREAMING_SNAKE_CASE :List[str] = feat_extract.model_input_names[0] __SCREAMING_SNAKE_CASE :Tuple = BatchFeature({input_name: speech_inputs} ) __SCREAMING_SNAKE_CASE :str = feat_extract.num_mel_bins # hack! __SCREAMING_SNAKE_CASE :Optional[int] = feat_extract.pad(lowercase_ ,padding='''longest''' ,return_tensors='''np''' )[input_name] __SCREAMING_SNAKE_CASE :Dict = feat_extract.pad(lowercase_ ,padding='''longest''' ,return_tensors='''pt''' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = self.feat_extract_dict __SCREAMING_SNAKE_CASE :str = True __SCREAMING_SNAKE_CASE :Optional[Any] = self.feature_extraction_class(**lowercase_ ) __SCREAMING_SNAKE_CASE :str = self.feat_extract_tester.prepare_inputs_for_target() __SCREAMING_SNAKE_CASE :Any = [len(lowercase_ ) for x in speech_inputs] __SCREAMING_SNAKE_CASE :List[str] = feat_extract.model_input_names[0] __SCREAMING_SNAKE_CASE :Any = BatchFeature({input_name: speech_inputs} ) __SCREAMING_SNAKE_CASE :Union[str, Any] = feat_extract.num_mel_bins # hack! __SCREAMING_SNAKE_CASE :Dict = feat_extract.pad(lowercase_ ,padding='''longest''' ,return_tensors='''np''' ) self.assertIn('''attention_mask''' ,lowercase_ ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,lowercase_ ) def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = self.feat_extract_dict __SCREAMING_SNAKE_CASE :Tuple = True __SCREAMING_SNAKE_CASE :int = self.feature_extraction_class(**lowercase_ ) __SCREAMING_SNAKE_CASE :int = self.feat_extract_tester.prepare_inputs_for_target() __SCREAMING_SNAKE_CASE :List[Any] = [len(lowercase_ ) for x in speech_inputs] __SCREAMING_SNAKE_CASE :Dict = feat_extract.model_input_names[0] __SCREAMING_SNAKE_CASE :List[str] = BatchFeature({input_name: speech_inputs} ) __SCREAMING_SNAKE_CASE :List[Any] = min(lowercase_ ) __SCREAMING_SNAKE_CASE :str = feat_extract.num_mel_bins # hack! __SCREAMING_SNAKE_CASE :Union[str, Any] = feat_extract.pad( lowercase_ ,padding='''max_length''' ,max_length=lowercase_ ,truncation=lowercase_ ,return_tensors='''np''' ) self.assertIn('''attention_mask''' ,lowercase_ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ) -> Dict: """simple docstring""" from datasets import load_dataset __SCREAMING_SNAKE_CASE :Tuple = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' ,'''clean''' ,split='''validation''' ) # automatic decoding with librispeech __SCREAMING_SNAKE_CASE :Optional[Any] = ds.sort('''id''' ).select(range(lowercase_ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :int = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on __SCREAMING_SNAKE_CASE :List[str] = self._load_datasamples(1 ) __SCREAMING_SNAKE_CASE :str = SpeechTaFeatureExtractor() __SCREAMING_SNAKE_CASE :int = feature_extractor(lowercase_ ,return_tensors='''pt''' ).input_values self.assertEquals(input_values.shape ,(1, 9_36_80) ) self.assertTrue(torch.allclose(input_values[0, :30] ,lowercase_ ,atol=1E-6 ) ) def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :int = torch.tensor( [-2.6_8_7_0, -3.0_1_0_4, -3.1_3_5_6, -3.5_3_5_2, -3.0_0_4_4, -3.0_3_5_3, -3.4_7_1_9, -3.6_7_7_7, -3.1_5_2_0, -2.9_4_3_5, -2.6_5_5_3, -2.8_7_9_5, -2.9_9_4_4, -2.5_9_2_1, -3.0_2_7_9, -3.0_3_8_6, -3.0_8_6_4, -3.1_2_9_1, -3.2_3_5_3, -2.7_4_4_4, -2.6_8_3_1, -2.7_2_8_7, -3.1_7_6_1, -3.1_5_7_1, -3.2_7_2_6, -3.0_5_8_2, -3.1_0_0_7, -3.4_5_3_3, -3.4_6_9_5, -3.0_9_9_8] ) # fmt: on __SCREAMING_SNAKE_CASE :str = self._load_datasamples(1 ) __SCREAMING_SNAKE_CASE :str = SpeechTaFeatureExtractor() __SCREAMING_SNAKE_CASE :int = feature_extractor(audio_target=lowercase_ ,return_tensors='''pt''' ).input_values self.assertEquals(input_values.shape ,(1, 3_66, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] ,lowercase_ ,atol=1E-4 ) )
371
"""simple docstring""" def __lowerCamelCase ( a_ : Union[str, Any] , a_ : Optional[Any] ) -> Union[str, Any]: return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def __lowerCamelCase ( a_ : Optional[int] , a_ : Any=0 ) -> Optional[Any]: return sorted(a_ , key=lambda a_ : x[column] ) def __lowerCamelCase ( a_ : Optional[Any] , a_ : Optional[int] , a_ : str=float('''inf''' ) ) -> str: for i in range(points_counts - 1 ): for j in range(i + 1 , a_ ): __SCREAMING_SNAKE_CASE :Dict = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __SCREAMING_SNAKE_CASE :Optional[Any] = current_dis return min_dis def __lowerCamelCase ( a_ : List[Any] , a_ : Any , a_ : Optional[int]=float('''inf''' ) ) -> Optional[Any]: for i in range(min(6 , points_counts - 1 ) , a_ ): for j in range(max(0 , i - 6 ) , a_ ): __SCREAMING_SNAKE_CASE :Dict = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __SCREAMING_SNAKE_CASE :int = current_dis return min_dis def __lowerCamelCase ( a_ : str , a_ : List[Any] , a_ : int ) -> Optional[int]: # base case if points_counts <= 3: return dis_between_closest_pair(a_ , a_ ) # recursion __SCREAMING_SNAKE_CASE :int = points_counts // 2 __SCREAMING_SNAKE_CASE :Dict = closest_pair_of_points_sqr( a_ , points_sorted_on_y[:mid] , a_ ) __SCREAMING_SNAKE_CASE :Any = closest_pair_of_points_sqr( a_ , points_sorted_on_y[mid:] , points_counts - mid ) __SCREAMING_SNAKE_CASE :Union[str, Any] = min(a_ , a_ ) __SCREAMING_SNAKE_CASE :str = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(a_ ) __SCREAMING_SNAKE_CASE :Dict = dis_between_closest_in_strip( a_ , len(a_ ) , a_ ) return min(a_ , a_ ) def __lowerCamelCase ( a_ : int , a_ : Tuple ) -> List[Any]: __SCREAMING_SNAKE_CASE :Union[str, Any] = column_based_sort(a_ , column=0 ) __SCREAMING_SNAKE_CASE :int = column_based_sort(a_ , column=1 ) return ( closest_pair_of_points_sqr( a_ , a_ , a_ ) ) ** 0.5 if __name__ == "__main__": lowerCamelCase_ = [(2, 3), (1_2, 3_0), (4_0, 5_0), (5, 1), (1_2, 1_0), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
239
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class a__ ( a_, unittest.TestCase ): __lowerCAmelCase = ShapEPipeline __lowerCAmelCase = ["""prompt"""] __lowerCAmelCase = ["""prompt"""] __lowerCAmelCase = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] __lowerCAmelCase = False @property def __magic_name__ ( self ): return 32 @property def __magic_name__ ( self ): return 32 @property def __magic_name__ ( self ): return self.time_input_dim * 4 @property def __magic_name__ ( self ): return 8 @property def __magic_name__ ( self ): lowercase : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def __magic_name__ ( self ): torch.manual_seed(0 ) lowercase : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModelWithProjection(_a ) @property def __magic_name__ ( self ): torch.manual_seed(0 ) lowercase : Any = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } lowercase : Tuple = PriorTransformer(**_a ) return model @property def __magic_name__ ( self ): torch.manual_seed(0 ) lowercase : Optional[int] = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } lowercase : Dict = ShapERenderer(**_a ) return model def __magic_name__ ( self ): lowercase : Optional[int] = self.dummy_prior lowercase : Optional[int] = self.dummy_text_encoder lowercase : Optional[int] = self.dummy_tokenizer lowercase : str = self.dummy_renderer lowercase : Dict = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1_024 , prediction_type="sample" , use_karras_sigmas=_a , clip_sample=_a , clip_sample_range=1.0 , ) lowercase : Any = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def __magic_name__ ( self , _a , _a=0 ): if str(_a ).startswith("mps" ): lowercase : str = torch.manual_seed(_a ) else: lowercase : List[str] = torch.Generator(device=_a ).manual_seed(_a ) lowercase : Tuple = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def __magic_name__ ( self ): lowercase : str = "cpu" lowercase : str = self.get_dummy_components() lowercase : Dict = self.pipeline_class(**_a ) lowercase : int = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowercase : List[str] = pipe(**self.get_dummy_inputs(_a ) ) lowercase : Optional[int] = output.images[0] lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) lowercase : Any = np.array( [ 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __magic_name__ ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __magic_name__ ( self ): lowercase : Tuple = torch_device == "cpu" lowercase : List[str] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_a , relax_max_difference=_a , ) def __magic_name__ ( self ): lowercase : int = self.get_dummy_components() lowercase : List[str] = self.pipeline_class(**_a ) lowercase : List[str] = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowercase : Optional[Any] = 1 lowercase : Tuple = 2 lowercase : Any = self.get_dummy_inputs(_a ) for key in inputs.keys(): if key in self.batch_params: lowercase : Any = batch_size * [inputs[key]] lowercase : Optional[int] = pipe(**_a , num_images_per_prompt=_a )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class a__ ( unittest.TestCase ): def __magic_name__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self ): lowercase : List[str] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy" ) lowercase : Tuple = ShapEPipeline.from_pretrained("openai/shap-e" ) lowercase : Tuple = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowercase : Optional[int] = torch.Generator(device=_a ).manual_seed(0 ) lowercase : Any = pipe( "a shark" , generator=_a , guidance_scale=1_5.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_a , _a )
202
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __magic_name__ ( __snake_case : Dict , __snake_case : Optional[Any]=False ) -> Tuple: lowercase : Union[str, Any] = OmegaConf.load(__snake_case ) if display: print(yaml.dump(OmegaConf.to_container(__snake_case ) ) ) return config def __magic_name__ ( __snake_case : Dict , __snake_case : Optional[Any]=None , __snake_case : Union[str, Any]=None ) -> Tuple: if conf_path is None: lowercase : List[Any] = "./model_checkpoints/vqgan_only.yaml" lowercase : Tuple = load_config(__snake_case , display=__snake_case ) lowercase : List[Any] = VQModel(**config.model.params ) if ckpt_path is None: lowercase : List[str] = "./model_checkpoints/vqgan_only.pt" lowercase : Optional[int] = torch.load(__snake_case , map_location=__snake_case ) if ".ckpt" in ckpt_path: lowercase : str = sd["state_dict"] model.load_state_dict(__snake_case , strict=__snake_case ) model.to(__snake_case ) del sd return model def __magic_name__ ( __snake_case : Tuple , __snake_case : Union[str, Any] ) -> int: lowercase , lowercase , lowercase : List[Any] = model.encode(__snake_case ) print(f"""VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}""" ) lowercase : str = model.decode(__snake_case ) return xrec def __magic_name__ ( __snake_case : Dict , __snake_case : Optional[int]=False ) -> int: lowercase , lowercase : Union[str, Any] = string.rsplit("." , 1 ) if reload: lowercase : Any = importlib.import_module(__snake_case ) importlib.reload(__snake_case ) return getattr(importlib.import_module(__snake_case , package=__snake_case ) , cls ) def __magic_name__ ( __snake_case : str ) -> List[str]: if "target" not in config: raise KeyError("Expected key `target` to instantiate." ) return get_obj_from_str(config["target"] )(**config.get("params" , {} ) ) def __magic_name__ ( __snake_case : Any , __snake_case : int , __snake_case : List[Any]=True , __snake_case : Dict=True ) -> str: lowercase : Optional[int] = instantiate_from_config(__snake_case ) if sd is not None: model.load_state_dict(__snake_case ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __magic_name__ ( __snake_case : Optional[int] , __snake_case : Any , __snake_case : Optional[int] , __snake_case : List[str] ) -> Any: # load the specified checkpoint if ckpt: lowercase : Dict = torch.load(__snake_case , map_location="cpu" ) lowercase : List[Any] = pl_sd["global_step"] print(f"""loaded model from global step {global_step}.""" ) else: lowercase : int = {"state_dict": None} lowercase : Optional[Any] = None lowercase : List[Any] = load_model_from_config(config.model , pl_sd["state_dict"] , gpu=__snake_case , eval_mode=__snake_case )["model"] return model, global_step
202
1
'''simple docstring''' def lowerCamelCase__ ( __lowerCamelCase : str ): return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
361
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase ={ 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'GroupViTTextConfig', 'GroupViTVisionConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase =[ 'GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GroupViTModel', 'GroupViTPreTrainedModel', 'GroupViTTextModel', 'GroupViTVisionModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase =[ 'TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFGroupViTModel', 'TFGroupViTPreTrainedModel', 'TFGroupViTTextModel', 'TFGroupViTVisionModel', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowercase =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
242
0
# Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() __UpperCAmelCase = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model __UpperCAmelCase = { # fairseq: '''wmt19-ru-en''': {'''length_penalty''': 1.1}, '''wmt19-en-ru''': {'''length_penalty''': 1.15}, '''wmt19-en-de''': {'''length_penalty''': 1.0}, '''wmt19-de-en''': {'''length_penalty''': 1.1}, # allenai: '''wmt16-en-de-dist-12-1''': {'''length_penalty''': 0.6}, '''wmt16-en-de-dist-6-1''': {'''length_penalty''': 0.6}, '''wmt16-en-de-12-1''': {'''length_penalty''': 0.8}, '''wmt19-de-en-6-6-base''': {'''length_penalty''': 0.6}, '''wmt19-de-en-6-6-big''': {'''length_penalty''': 0.6}, } # this remaps the different models to their organization names __UpperCAmelCase = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __UpperCAmelCase = '''facebook''' for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: __UpperCAmelCase = '''allenai''' def UpperCamelCase ( snake_case__ : int ) -> Optional[Any]: # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} UpperCamelCase : Dict = dict((re.sub(R'@@$' , '' , snake_case__ ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , snake_case__ ), v) for k, v in d.items() ) UpperCamelCase : List[Any] = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] UpperCamelCase : str = d[k] # restore return da def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Tuple ) -> List[Any]: # prep assert os.path.exists(snake_case__ ) os.makedirs(snake_case__ , exist_ok=snake_case__ ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models UpperCamelCase : Any = basename(snake_case__ ) UpperCamelCase : Optional[int] = dirname(snake_case__ ) UpperCamelCase : Tuple = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel UpperCamelCase : int = cls.hub_models() UpperCamelCase : Dict = {'bpe': 'fastbpe', 'tokenizer': 'moses'} UpperCamelCase : Tuple = '.' # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(F"""using checkpoint {checkpoint_file}""" ) UpperCamelCase : Tuple = hub_utils.from_pretrained( snake_case__ , snake_case__ , snake_case__ , archive_map=snake_case__ , **snake_case__ ) UpperCamelCase : Union[str, Any] = vars(chkpt['args']['model'] ) UpperCamelCase : Optional[Any] = args['source_lang'] UpperCamelCase : int = args['target_lang'] UpperCamelCase : Dict = dirname(snake_case__ ) UpperCamelCase : List[Any] = basename(snake_case__ ) # dicts UpperCamelCase : Dict = os.path.join(snake_case__ , F"""dict.{src_lang}.txt""" ) UpperCamelCase : Optional[Any] = os.path.join(snake_case__ , F"""dict.{tgt_lang}.txt""" ) UpperCamelCase : Optional[Any] = Dictionary.load(snake_case__ ) UpperCamelCase : str = rewrite_dict_keys(src_dict.indices ) UpperCamelCase : List[Any] = len(snake_case__ ) UpperCamelCase : Dict = os.path.join(snake_case__ , 'vocab-src.json' ) print(F"""Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , ensure_ascii=snake_case__ , indent=snake_case__ ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab UpperCamelCase : Optional[Any] = True for k in src_vocab.keys(): if not k.islower(): UpperCamelCase : List[Any] = False break UpperCamelCase : Optional[Any] = Dictionary.load(snake_case__ ) UpperCamelCase : Optional[int] = rewrite_dict_keys(tgt_dict.indices ) UpperCamelCase : Optional[int] = len(snake_case__ ) UpperCamelCase : Tuple = os.path.join(snake_case__ , 'vocab-tgt.json' ) print(F"""Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , ensure_ascii=snake_case__ , indent=snake_case__ ) ) # merges_file (bpecodes) UpperCamelCase : Dict = os.path.join(snake_case__ , VOCAB_FILES_NAMES['merges_file'] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" UpperCamelCase : Optional[int] = os.path.join(snake_case__ , snake_case__ ) if os.path.exists(snake_case__ ): break with open(snake_case__ , encoding='utf-8' ) as fin: UpperCamelCase : List[str] = fin.read() UpperCamelCase : List[Any] = re.sub(R' \d+$' , '' , snake_case__ , 0 , re.M ) # remove frequency number print(F"""Generating {merges_file}""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as fout: fout.write(snake_case__ ) # model config UpperCamelCase : List[Any] = os.path.join(snake_case__ , 'config.json' ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", F"""need to extend tokenizer to support bpe={args["bpe"]}""" assert args["tokenizer"] == "moses", F"""need to extend tokenizer to support bpe={args["tokenizer"]}""" UpperCamelCase : Optional[Any] = { 'architectures': ['FSMTForConditionalGeneration'], 'model_type': 'fsmt', 'activation_dropout': args['activation_dropout'], 'activation_function': 'relu', 'attention_dropout': args['attention_dropout'], 'd_model': args['decoder_embed_dim'], 'dropout': args['dropout'], 'init_std': 0.02, 'max_position_embeddings': args['max_source_positions'], 'num_hidden_layers': args['encoder_layers'], 'src_vocab_size': src_vocab_size, 'tgt_vocab_size': tgt_vocab_size, 'langs': [src_lang, tgt_lang], 'encoder_attention_heads': args['encoder_attention_heads'], 'encoder_ffn_dim': args['encoder_ffn_embed_dim'], 'encoder_layerdrop': args['encoder_layerdrop'], 'encoder_layers': args['encoder_layers'], 'decoder_attention_heads': args['decoder_attention_heads'], 'decoder_ffn_dim': args['decoder_ffn_embed_dim'], 'decoder_layerdrop': args['decoder_layerdrop'], 'decoder_layers': args['decoder_layers'], 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'is_encoder_decoder': True, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_all_embeddings'], } # good hparam defaults to start with UpperCamelCase : str = 5 UpperCamelCase : Tuple = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: UpperCamelCase : List[Any] = best_score_hparams[model_dir]['length_penalty'] else: UpperCamelCase : int = 1.0 print(F"""Generating {fsmt_model_config_file}""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , ensure_ascii=snake_case__ , indent=snake_case__ ) ) # tokenizer config UpperCamelCase : Tuple = os.path.join(snake_case__ , snake_case__ ) UpperCamelCase : Optional[Any] = { 'langs': [src_lang, tgt_lang], 'model_max_length': 1024, 'do_lower_case': do_lower_case, } print(F"""Generating {fsmt_tokenizer_config_file}""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , ensure_ascii=snake_case__ , indent=snake_case__ ) ) # model UpperCamelCase : List[Any] = chkpt['models'][0] UpperCamelCase : Optional[int] = model.state_dict() # rename keys to start with 'model.' UpperCamelCase : Optional[Any] = OrderedDict(('model.' + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys UpperCamelCase : Tuple = [ 'model.model', 'model.encoder.version', 'model.decoder.version', 'model.encoder_embed_tokens.weight', 'model.decoder_embed_tokens.weight', 'model.encoder.embed_positions._float_tensor', 'model.decoder.embed_positions._float_tensor', ] for k in ignore_keys: model_state_dict.pop(snake_case__ , snake_case__ ) UpperCamelCase : Union[str, Any] = FSMTConfig.from_pretrained(snake_case__ ) UpperCamelCase : List[str] = FSMTForConditionalGeneration(snake_case__ ) # check that it loads ok model_new.load_state_dict(snake_case__ , strict=snake_case__ ) # save UpperCamelCase : List[str] = os.path.join(snake_case__ , snake_case__ ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(snake_case__ , snake_case__ ) print('Conversion is done!' ) print('\nLast step is to upload the files to s3' ) print(F"""cd {data_root}""" ) print(F"""transformers-cli upload {model_dir}""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fsmt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __UpperCAmelCase = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
119
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu __UpperCAmelCase = get_tests_dir() + '''/test_data/fsmt/fsmt_val_data.json''' with io.open(filename, '''r''', encoding='''utf-8''') as f: __UpperCAmelCase = json.load(f) @require_torch class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: return FSMTTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: UpperCamelCase : List[Any] = FSMTForConditionalGeneration.from_pretrained(SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['en-ru', 26.0], ['ru-en', 22.0], ['en-de', 22.0], ['de-en', 29.0], ] ) @slow def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality UpperCamelCase : int = F"""facebook/wmt19-{pair}""" UpperCamelCase : Optional[int] = self.get_tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : str = self.get_model(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Dict = bleu_data[pair]['src'] UpperCamelCase : Tuple = bleu_data[pair]['tgt'] UpperCamelCase : List[Any] = tokenizer(SCREAMING_SNAKE_CASE_, return_tensors='pt', truncation=SCREAMING_SNAKE_CASE_, padding='longest' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Any = model.generate( input_ids=batch.input_ids, num_beams=8, ) UpperCamelCase : Tuple = tokenizer.batch_decode( SCREAMING_SNAKE_CASE_, skip_special_tokens=SCREAMING_SNAKE_CASE_, clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[Any] = calculate_bleu(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) print(SCREAMING_SNAKE_CASE_ ) self.assertGreaterEqual(scores['bleu'], SCREAMING_SNAKE_CASE_ )
119
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''google/bigbird-roberta-base''': '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json''', '''google/bigbird-roberta-large''': '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json''', '''google/bigbird-base-trivia-itc''': '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json''', # See all BigBird models at https://huggingface.co/models?filter=big_bird } class A ( __UpperCAmelCase ): __snake_case = 'big_bird' def __init__( self, UpperCamelCase__=5_0358, UpperCamelCase__=768, UpperCamelCase__=12, UpperCamelCase__=12, UpperCamelCase__=3072, UpperCamelCase__="gelu_new", UpperCamelCase__=0.1, UpperCamelCase__=0.1, UpperCamelCase__=4096, UpperCamelCase__=2, UpperCamelCase__=0.02, UpperCamelCase__=1E-12, UpperCamelCase__=True, UpperCamelCase__=0, UpperCamelCase__=1, UpperCamelCase__=2, UpperCamelCase__=66, UpperCamelCase__="block_sparse", UpperCamelCase__=True, UpperCamelCase__=False, UpperCamelCase__=64, UpperCamelCase__=3, UpperCamelCase__=None, **UpperCamelCase__, ): """simple docstring""" super().__init__( pad_token_id=UpperCamelCase__, bos_token_id=UpperCamelCase__, eos_token_id=UpperCamelCase__, sep_token_id=UpperCamelCase__, **UpperCamelCase__, ) lowerCAmelCase_ = vocab_size lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_act lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = initializer_range lowerCAmelCase_ = type_vocab_size lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = use_cache lowerCAmelCase_ = rescale_embeddings lowerCAmelCase_ = attention_type lowerCAmelCase_ = use_bias lowerCAmelCase_ = block_size lowerCAmelCase_ = num_random_blocks lowerCAmelCase_ = classifier_dropout class A ( __UpperCAmelCase ): @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" if self.task == "multiple-choice": lowerCAmelCase_ = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase_ = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
167
import string def __UpperCamelCase ( _A ): for key in range(len(string.ascii_uppercase ) ): lowerCAmelCase_ = '''''' for symbol in message: if symbol in string.ascii_uppercase: lowerCAmelCase_ = string.ascii_uppercase.find(_A ) lowerCAmelCase_ = num - key if num < 0: lowerCAmelCase_ = num + len(string.ascii_uppercase ) lowerCAmelCase_ = translated + string.ascii_uppercase[num] else: lowerCAmelCase_ = translated + symbol print(f"Decryption using Key #{key}: {translated}" ) def __UpperCamelCase ( ): lowerCAmelCase_ = input('''Encrypted message: ''' ) lowerCAmelCase_ = message.upper() decrypt(_A ) if __name__ == "__main__": import doctest doctest.testmod() main()
167
1
"""simple docstring""" import requests from bsa import BeautifulSoup def _SCREAMING_SNAKE_CASE ( __snake_case : str = "AAPL" ): '''simple docstring''' lowercase = f'https://in.finance.yahoo.com/quote/{symbol}?s={symbol}' lowercase = BeautifulSoup(requests.get(__snake_case ).text , 'html.parser' ) lowercase = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(F'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
220
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( __snake_case : int = 4_00_00_00 ): '''simple docstring''' lowercase = [] lowercase , lowercase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__snake_case ) lowercase , lowercase = b, a + b return sum(__snake_case ) if __name__ == "__main__": print(F'''{solution() = }''')
220
1
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class lowercase__ ( unittest.TestCase ): A__ : Union[str, Any] =MODEL_FOR_MASKED_LM_MAPPING A__ : Dict =TF_MODEL_FOR_MASKED_LM_MAPPING def A_ ( self : List[str] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def A_ ( self : Dict ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='tf' ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ {'sequence': 'My name is grouped', 'score': 2.1e-0_5, 'token': 38015, 'token_str': ' grouped'}, {'sequence': 'My name is accuser', 'score': 2.1e-0_5, 'token': 25506, 'token_str': ' accuser'}, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ { 'sequence': 'The largest city in France is grouped', 'score': 2.1e-0_5, 'token': 38015, 'token_str': ' grouped', }, { 'sequence': 'The largest city in France is accuser', 'score': 2.1e-0_5, 'token': 25506, 'token_str': ' accuser', }, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ {'sequence': 'My name is Clara', 'score': 2e-0_5, 'token': 13606, 'token_str': ' Clara'}, {'sequence': 'My name is Patrick', 'score': 2e-0_5, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 1.9e-0_5, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def A_ ( self : int ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='pt' ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ {'sequence': 'My name is Maul', 'score': 2.2e-0_5, 'token': 35676, 'token_str': ' Maul'}, {'sequence': 'My name isELS', 'score': 2.2e-0_5, 'token': 16416, 'token_str': 'ELS'}, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ { 'sequence': 'The largest city in France is Maul', 'score': 2.2e-0_5, 'token': 35676, 'token_str': ' Maul', }, {'sequence': 'The largest city in France isELS', 'score': 2.2e-0_5, 'token': 16416, 'token_str': 'ELS'}, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ {'sequence': 'My name is Patrick', 'score': 2.1e-0_5, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 2e-0_5, 'token': 2941, 'token_str': ' Te'}, {'sequence': 'My name is Clara', 'score': 2e-0_5, 'token': 13606, 'token_str': ' Clara'}, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask> <mask>' , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=6 ) , [ [ { 'score': 2.2e-0_5, 'token': 35676, 'token_str': ' Maul', 'sequence': '<s>My name is Maul<mask></s>', }, {'score': 2.2e-0_5, 'token': 16416, 'token_str': 'ELS', 'sequence': '<s>My name isELS<mask></s>'}, ], [ { 'score': 2.2e-0_5, 'token': 35676, 'token_str': ' Maul', 'sequence': '<s>My name is<mask> Maul</s>', }, {'score': 2.2e-0_5, 'token': 16416, 'token_str': 'ELS', 'sequence': '<s>My name is<mask>ELS</s>'}, ], ] , ) @require_torch_gpu def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = pipeline('fill-mask' , model='hf-internal-testing/tiny-random-distilbert' , device=0 , framework='pt' ) # convert model to fp16 pipe.model.half() SCREAMING_SNAKE_CASE__ = pipe('Paris is the [MASK] of France.' ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) @slow @require_torch def A_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='pt' ) self.run_large_test(UpperCAmelCase_ ) @slow @require_tf def A_ ( self : Tuple ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='tf' ) self.run_large_test(UpperCAmelCase_ ) def A_ ( self : List[str] , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ {'sequence': 'My name is John', 'score': 0.008, 'token': 610, 'token_str': ' John'}, {'sequence': 'My name is Chris', 'score': 0.007, 'token': 1573, 'token_str': ' Chris'}, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ { 'sequence': 'The largest city in France is Paris', 'score': 0.251, 'token': 2201, 'token_str': ' Paris', }, { 'sequence': 'The largest city in France is Lyon', 'score': 0.214, 'token': 12790, 'token_str': ' Lyon', }, ] , ) SCREAMING_SNAKE_CASE__ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ {'sequence': 'My name is Patrick', 'score': 0.005, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Clara', 'score': 0.000, 'token': 13606, 'token_str': ' Clara'}, {'sequence': 'My name is Te', 'score': 0.000, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def A_ ( self : List[str] ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='pt' ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None self.run_pipeline_test(UpperCAmelCase_ , [] ) @require_tf def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='tf' ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None self.run_pipeline_test(UpperCAmelCase_ , [] ) def A_ ( self : List[str] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : int ): if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest('The provided tokenizer has no mask token, (probably reformer or wav2vec2)' ) SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = [ F'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def A_ ( self : Tuple , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[str] ): SCREAMING_SNAKE_CASE__ = fill_masker.tokenizer SCREAMING_SNAKE_CASE__ = fill_masker.model SCREAMING_SNAKE_CASE__ = fill_masker( F'This is a {tokenizer.mask_token}' , ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) SCREAMING_SNAKE_CASE__ = fill_masker([F'This is a {tokenizer.mask_token}'] ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) SCREAMING_SNAKE_CASE__ = fill_masker([F'This is a {tokenizer.mask_token}', F'Another {tokenizer.mask_token} great test.'] ) self.assertEqual( UpperCAmelCase_ , [ [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ], [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ], ] , ) with self.assertRaises(UpperCAmelCase_ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(UpperCAmelCase_ ): fill_masker('This is' ) self.run_test_top_k(UpperCAmelCase_ , UpperCAmelCase_ ) self.run_test_targets(UpperCAmelCase_ , UpperCAmelCase_ ) self.run_test_top_k_targets(UpperCAmelCase_ , UpperCAmelCase_ ) self.fill_mask_with_duplicate_targets_and_top_k(UpperCAmelCase_ , UpperCAmelCase_ ) self.fill_mask_with_multiple_masks(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE__ = tokenizer.get_vocab() SCREAMING_SNAKE_CASE__ = sorted(vocab.keys() )[:2] # Pipeline argument SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ , targets=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) SCREAMING_SNAKE_CASE__ = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(UpperCAmelCase_ ) ) # Call argument SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets=UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) SCREAMING_SNAKE_CASE__ = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(UpperCAmelCase_ ) ) # Score equivalence SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = [top_mask['token_str'] for top_mask in outputs] SCREAMING_SNAKE_CASE__ = [top_mask['score'] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(UpperCAmelCase_ ) == set(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = [top_mask['score'] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(UpperCAmelCase_ ) , nested_simplify(UpperCAmelCase_ ) ) # Raises with invalid with self.assertRaises(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets=[''] ) with self.assertRaises(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , targets='' ) def A_ ( self : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : str ): SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ , top_k=2 ) SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=2 ) self.assertEqual( UpperCAmelCase_ , [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ] , ) self.assertEqual(nested_simplify(UpperCAmelCase_ ) , nested_simplify(UpperCAmelCase_ ) ) def A_ ( self : Union[str, Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE__ = tokenizer.get_vocab() SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) # top_k=2, ntargets=3 SCREAMING_SNAKE_CASE__ = sorted(vocab.keys() )[:3] SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=2 , targets=UpperCAmelCase_ ) # If we use the most probably targets, and filter differently, we should still # have the same results SCREAMING_SNAKE_CASE__ = [el['token_str'] for el in sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : x["score"] , reverse=UpperCAmelCase_ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(UpperCAmelCase_ ).issubset(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=3 , targets=UpperCAmelCase_ ) # They should yield exactly the same result self.assertEqual(nested_simplify(UpperCAmelCase_ ) , nested_simplify(UpperCAmelCase_ ) ) def A_ ( self : Dict , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str ): SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.get_vocab() # String duplicates + id duplicates SCREAMING_SNAKE_CASE__ = sorted(vocab.keys() )[:3] SCREAMING_SNAKE_CASE__ = [targets[0], targets[1], targets[0], targets[2], targets[1]] SCREAMING_SNAKE_CASE__ = fill_masker(F'My name is {tokenizer.mask_token}' , targets=UpperCAmelCase_ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(UpperCAmelCase_ ) , 3 ) def A_ ( self : Optional[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE__ = FillMaskPipeline(model=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = fill_masker( F'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2 ) self.assertEqual( UpperCAmelCase_ , [ [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ], [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ], [ {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, {'sequence': ANY(UpperCAmelCase_ ), 'score': ANY(UpperCAmelCase_ ), 'token': ANY(UpperCAmelCase_ ), 'token_str': ANY(UpperCAmelCase_ )}, ], ] , )
169
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class lowercase__ : @staticmethod def A_ ( *UpperCAmelCase_ : Optional[Any] , **UpperCAmelCase_ : Optional[Any] ): pass @is_pipeline_test @require_vision class lowercase__ ( unittest.TestCase ): @require_torch def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , ) SCREAMING_SNAKE_CASE__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) SCREAMING_SNAKE_CASE__ = image_classifier(UpperCAmelCase_ , candidate_labels=['a', 'b', 'c'] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(UpperCAmelCase_ ) , [ [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'b'}, {'score': 0.333, 'label': 'c'}], [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'c'}, {'score': 0.333, 'label': 'b'}], ] , ) SCREAMING_SNAKE_CASE__ = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], ] , ) @require_tf def A_ ( self : List[Any] ): SCREAMING_SNAKE_CASE__ = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , framework='tf' ) SCREAMING_SNAKE_CASE__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) SCREAMING_SNAKE_CASE__ = image_classifier(UpperCAmelCase_ , candidate_labels=['a', 'b', 'c'] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'b'}, {'score': 0.333, 'label': 'c'}] , ) SCREAMING_SNAKE_CASE__ = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], [ {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, {'score': 0.333, 'label': ANY(UpperCAmelCase_ )}, ], ] , ) @slow @require_torch def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , ) # This is an image of 2 cats with remotes and no planes SCREAMING_SNAKE_CASE__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) SCREAMING_SNAKE_CASE__ = image_classifier(UpperCAmelCase_ , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ] , ) SCREAMING_SNAKE_CASE__ = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ], ] * 5 , ) @slow @require_tf def A_ ( self : Tuple ): SCREAMING_SNAKE_CASE__ = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , framework='tf' ) # This is an image of 2 cats with remotes and no planes SCREAMING_SNAKE_CASE__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) SCREAMING_SNAKE_CASE__ = image_classifier(UpperCAmelCase_ , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ] , ) SCREAMING_SNAKE_CASE__ = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ) , [ [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ], ] * 5 , )
169
1
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class A__ : """simple docstring""" def __init__( self , lowercase , lowercase=3 , lowercase=32 , lowercase=3 , lowercase=10 , lowercase=[8, 16, 32, 64] , lowercase=[1, 1, 2, 1] , lowercase=True , lowercase=True , lowercase="relu" , lowercase=3 , lowercase=None , lowercase=["stage2", "stage3", "stage4"] , lowercase=[2, 3, 4] , lowercase=1 , ) -> Optional[int]: '''simple docstring''' a__ : Optional[Any] = parent a__ : Any = batch_size a__ : Any = image_size a__ : Union[str, Any] = num_channels a__ : Optional[int] = embeddings_size a__ : str = hidden_sizes a__ : Optional[Any] = depths a__ : Optional[Any] = is_training a__ : int = use_labels a__ : Union[str, Any] = hidden_act a__ : Optional[int] = num_labels a__ : Optional[int] = scope a__ : Tuple = len(lowercase) a__ : Union[str, Any] = out_features a__ : List[str] = out_indices a__ : Any = num_groups def __lowercase ( self) -> str: '''simple docstring''' a__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) a__ : Optional[Any] = None if self.use_labels: a__ : Tuple = ids_tensor([self.batch_size] , self.num_labels) a__ : Optional[int] = self.get_config() return config, pixel_values, labels def __lowercase ( self) -> str: '''simple docstring''' return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def __lowercase ( self , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' a__ : List[Any] = BitModel(config=lowercase) model.to(lowercase) model.eval() a__ : int = model(lowercase) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __lowercase ( self , lowercase , lowercase , lowercase) -> str: '''simple docstring''' a__ : Union[str, Any] = self.num_labels a__ : Union[str, Any] = BitForImageClassification(lowercase) model.to(lowercase) model.eval() a__ : List[Any] = model(lowercase , labels=lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __lowercase ( self , lowercase , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ : List[Any] = BitBackbone(config=lowercase) model.to(lowercase) model.eval() a__ : List[str] = model(lowercase) # verify feature maps self.parent.assertEqual(len(result.feature_maps) , len(config.out_features)) self.parent.assertListEqual(list(result.feature_maps[0].shape) , [self.batch_size, self.hidden_sizes[1], 4, 4]) # verify channels self.parent.assertEqual(len(model.channels) , len(config.out_features)) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:]) # verify backbone works with out_features=None a__ : Union[str, Any] = None a__ : List[Any] = BitBackbone(config=lowercase) model.to(lowercase) model.eval() a__ : Optional[int] = model(lowercase) # verify feature maps self.parent.assertEqual(len(result.feature_maps) , 1) self.parent.assertListEqual(list(result.feature_maps[0].shape) , [self.batch_size, self.hidden_sizes[-1], 1, 1]) # verify channels self.parent.assertEqual(len(model.channels) , 1) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]]) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : int = self.prepare_config_and_inputs() a__ , a__ , a__ : Optional[int] = config_and_inputs a__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" __A : Union[str, Any] = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __A : List[Any] = ( {'''feature-extraction''': BitModel, '''image-classification''': BitForImageClassification} if is_torch_available() else {} ) __A : Optional[Any] = False __A : int = False __A : Optional[int] = False __A : str = False __A : Optional[Any] = False def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : List[str] = BitModelTester(self) a__ : int = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase) def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' return @unittest.skip(reason='Bit does not output attentions') def __lowercase ( self) -> int: '''simple docstring''' pass @unittest.skip(reason='Bit does not use inputs_embeds') def __lowercase ( self) -> List[str]: '''simple docstring''' pass @unittest.skip(reason='Bit does not support input and output embeddings') def __lowercase ( self) -> Optional[int]: '''simple docstring''' pass def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ , a__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : str = model_class(lowercase) a__ : Optional[int] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__ : Optional[int] = [*signature.parameters.keys()] a__ : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase) def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase) def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ , a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : Union[str, Any] = model_class(config=lowercase) for name, module in model.named_modules(): if isinstance(lowercase , (nn.BatchNormad, nn.GroupNorm)): self.assertTrue( torch.all(module.weight == 1) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) def __lowercase ( self) -> Optional[Any]: '''simple docstring''' def check_hidden_states_output(lowercase , lowercase , lowercase): a__ : Optional[Any] = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__ : Union[str, Any] = model(**self._prepare_for_class(lowercase , lowercase)) a__ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states a__ : List[str] = self.model_tester.num_stages self.assertEqual(len(lowercase) , expected_num_stages + 1) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) a__ , a__ : Any = self.model_tester.prepare_config_and_inputs_for_common() a__ : int = ['preactivation', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: a__ : int = layer_type a__ : List[str] = True check_hidden_states_output(lowercase , lowercase , lowercase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a__ : int = True check_hidden_states_output(lowercase , lowercase , lowercase) @unittest.skip(reason='Bit does not use feedforward chunking') def __lowercase ( self) -> List[Any]: '''simple docstring''' pass def __lowercase ( self) -> str: '''simple docstring''' a__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase) @slow def __lowercase ( self) -> Optional[int]: '''simple docstring''' for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : Dict = BitModel.from_pretrained(lowercase) self.assertIsNotNone(lowercase) def A_ ( ) -> int: a__ : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self) -> Dict: '''simple docstring''' return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def __lowercase ( self) -> str: '''simple docstring''' a__ : Any = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]).to(lowercase) a__ : Tuple = self.default_image_processor a__ : str = prepare_img() a__ : Tuple = image_processor(images=lowercase , return_tensors='pt').to(lowercase) # forward pass with torch.no_grad(): a__ : Any = model(**lowercase) # verify the logits a__ : Any = torch.Size((1, 1000)) self.assertEqual(outputs.logits.shape , lowercase) a__ : Optional[Any] = torch.tensor([[-0.65_26, -0.52_63, -1.43_98]]).to(lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1e-4)) @require_torch class A__ ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" __A : Optional[int] = (BitBackbone,) if is_torch_available() else () __A : Optional[Any] = BitConfig __A : int = False def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : str = BitModelTester(self)
99
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params lowercase__ : Any = getLogger(__name__) lowercase__ : List[str] = "cuda" if torch.cuda.is_available() else "cpu" def A_ ( snake_case : List[str] , snake_case : str , snake_case : str , snake_case : int = 8 , snake_case : str = DEFAULT_DEVICE , snake_case : List[str]=False , snake_case : Union[str, Any]="summarization" , snake_case : str=None , **snake_case : List[Any] , ) -> Dict: '''simple docstring''' __UpperCamelCase = Path(snake_case ).open('''w''' , encoding='''utf-8''' ) __UpperCamelCase = str(snake_case ) __UpperCamelCase = AutoModelForSeqaSeqLM.from_pretrained(snake_case ).to(snake_case ) if fpaa: __UpperCamelCase = model.half() __UpperCamelCase = AutoTokenizer.from_pretrained(snake_case ) logger.info(f"Inferred tokenizer type: {tokenizer.__class__}" ) # if this is wrong, check config.model_type. __UpperCamelCase = time.time() # update config with task specific params use_task_specific_params(snake_case , snake_case ) if prefix is None: __UpperCamelCase = prefix or getattr(model.config , '''prefix''' , '''''' ) or '''''' for examples_chunk in tqdm(list(chunks(snake_case , snake_case ) ) ): __UpperCamelCase = [prefix + text for text in examples_chunk] __UpperCamelCase = tokenizer(snake_case , return_tensors='''pt''' , truncation=snake_case , padding='''longest''' ).to(snake_case ) __UpperCamelCase = model.generate( input_ids=batch.input_ids , attention_mask=batch.attention_mask , **snake_case , ) __UpperCamelCase = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case , clean_up_tokenization_spaces=snake_case ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() __UpperCamelCase = int(time.time() - start_time ) # seconds __UpperCamelCase = len(snake_case ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs , 4 )} def A_ ( ) -> Tuple: '''simple docstring''' return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def A_ ( snake_case : str=True ) -> int: '''simple docstring''' __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' , type=snake_case , help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' , type=snake_case , help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' , type=snake_case , help='''where to save summaries''' ) parser.add_argument('''--reference_path''' , type=snake_case , required=snake_case , help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' , type=snake_case , required=snake_case , default='''metrics.json''' , help='''where to save metrics''' ) parser.add_argument('''--device''' , type=snake_case , required=snake_case , default=snake_case , help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' , type=snake_case , required=snake_case , default=snake_case , help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' , type=snake_case , default='''summarization''' , help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' , type=snake_case , default=8 , required=snake_case , help='''batch size''' ) parser.add_argument( '''--n_obs''' , type=snake_case , default=-1 , required=snake_case , help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' , action='''store_true''' ) parser.add_argument('''--dump-args''' , action='''store_true''' , help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' , nargs='''?''' , type=snake_case , const=datetime_now() , help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) , ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __UpperCamelCase , __UpperCamelCase = parser.parse_known_args() __UpperCamelCase = parse_numeric_n_bool_cl_kwargs(snake_case ) if parsed_args and verbose: print(f"parsed the following generate kwargs: {parsed_args}" ) __UpperCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __UpperCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=snake_case ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(f"score_path {args.score_path} will be overwritten unless you type ctrl-c." ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) __UpperCamelCase = generate_summaries_or_translations( snake_case , args.save_path , args.model_name , batch_size=args.bs , device=args.device , fpaa=args.fpaa , task=args.task , prefix=args.prefix , **snake_case , ) if args.reference_path is None: return {} # Compute scores __UpperCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge __UpperCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] __UpperCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(snake_case )] __UpperCamelCase = score_fn(snake_case , snake_case ) scores.update(snake_case ) if args.dump_args: scores.update(snake_case ) if args.info: __UpperCamelCase = args.info if verbose: print(snake_case ) if args.score_path is not None: json.dump(snake_case , open(args.score_path , '''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
328
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, is_batched, to_numpy_array, valid_images, ) from ...utils import TensorType, logging lowerCAmelCase__ = logging.get_logger(__name__) class lowercase_ (_SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ['pixel_values'] def __init__( self : List[str] ,lowercase__ : Tuple = True ,lowercase__ : Optional[int] = None ,lowercase__ : Any = PILImageResampling.BICUBIC ,lowercase__ : Optional[int] = True ,lowercase__ : Union[str, Any] = True ,lowercase__ : Union[str, Any] = 1 / 2_5_5 ,lowercase__ : List[str] = None ,lowercase__ : Any = True ,lowercase__ : str = None ,lowercase__ : str = None ,**lowercase__ : List[str] ,): super().__init__(**lowercase__ ) __lowercase = size if size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} __lowercase = get_size_dict(lowercase__ ) __lowercase = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} __lowercase = get_size_dict(lowercase__ ,default_to_square=lowercase__ ,param_name='''crop_size''' ) __lowercase = do_resize __lowercase = do_rescale __lowercase = do_normalize __lowercase = do_center_crop __lowercase = crop_size __lowercase = size __lowercase = resample __lowercase = rescale_factor __lowercase = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __lowercase = image_std if image_std is not None else IMAGENET_DEFAULT_STD def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Any = PILImageResampling.BILINEAR ,lowercase__ : Optional[int] = None ,**lowercase__ : str ,): __lowercase = get_size_dict(lowercase__ ) if "shortest_edge" in size: __lowercase = get_resize_output_image_size(lowercase__ ,size=size['''shortest_edge'''] ,default_to_square=lowercase__ ) # size = get_resize_output_image_size(image, size["shortest_edge"], size["longest_edge"]) elif "height" in size and "width" in size: __lowercase = (size['''height'''], size['''width''']) else: raise ValueError(F"Size must contain \'height\' and \'width\' keys or \'shortest_edge\' key. Got {size.keys()}" ) return resize(lowercase__ ,size=lowercase__ ,resample=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Union[str, Any] ,lowercase__ : Any ,lowercase__ : str = None ,**lowercase__ : Tuple ,): __lowercase = get_size_dict(lowercase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowercase__ ,size=(size['''height'''], size['''width''']) ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[Any] = None ,**lowercase__ : Dict ): return rescale(lowercase__ ,scale=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[str] = None ,**lowercase__ : Any ,): return normalize(lowercase__ ,mean=lowercase__ ,std=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Dict ,lowercase__ : Any = None ,lowercase__ : Union[str, Any] = None ,lowercase__ : Any = None ,lowercase__ : Tuple = None ,lowercase__ : Optional[int] = None ,lowercase__ : Union[str, Any] = None ,lowercase__ : Optional[Any] = None ,lowercase__ : int = None ,lowercase__ : str = None ,lowercase__ : Tuple = None ,lowercase__ : Optional[int] = None ,lowercase__ : Optional[int] = ChannelDimension.FIRST ,**lowercase__ : Union[str, Any] ,): __lowercase = do_resize if do_resize is not None else self.do_resize __lowercase = do_rescale if do_rescale is not None else self.do_rescale __lowercase = do_normalize if do_normalize is not None else self.do_normalize __lowercase = do_center_crop if do_center_crop is not None else self.do_center_crop __lowercase = crop_size if crop_size is not None else self.crop_size __lowercase = get_size_dict(lowercase__ ,param_name='''crop_size''' ,default_to_square=lowercase__ ) __lowercase = resample if resample is not None else self.resample __lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase = image_mean if image_mean is not None else self.image_mean __lowercase = image_std if image_std is not None else self.image_std __lowercase = size if size is not None else self.size __lowercase = get_size_dict(lowercase__ ) if not is_batched(lowercase__ ): __lowercase = [images] if not valid_images(lowercase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. __lowercase = [to_numpy_array(lowercase__ ) for image in images] if do_resize: __lowercase = [self.resize(image=lowercase__ ,size=lowercase__ ,resample=lowercase__ ) for image in images] if do_center_crop: __lowercase = [self.center_crop(image=lowercase__ ,size=lowercase__ ) for image in images] if do_rescale: __lowercase = [self.rescale(image=lowercase__ ,scale=lowercase__ ) for image in images] if do_normalize: __lowercase = [self.normalize(image=lowercase__ ,mean=lowercase__ ,std=lowercase__ ) for image in images] __lowercase = [to_channel_dimension_format(lowercase__ ,lowercase__ ) for image in images] __lowercase = {'''pixel_values''': images} return BatchFeature(data=lowercase__ ,tensor_type=lowercase__ )
365
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : List[str] ,lowercase__ : str ,lowercase__ : List[str]=1_3 ,lowercase__ : str=7 ,lowercase__ : Tuple=True ,lowercase__ : Dict=True ,lowercase__ : str=True ,lowercase__ : Optional[Any]=True ,lowercase__ : List[str]=9_9 ,lowercase__ : int=[1, 1, 2] ,lowercase__ : int=1 ,lowercase__ : Tuple=3_2 ,lowercase__ : Union[str, Any]=4 ,lowercase__ : Tuple=8 ,lowercase__ : Any=3_7 ,lowercase__ : Union[str, Any]="gelu_new" ,lowercase__ : Tuple=0.1 ,lowercase__ : int=0.1 ,lowercase__ : Optional[int]=0.0 ,lowercase__ : Union[str, Any]=5_1_2 ,lowercase__ : Dict=3 ,lowercase__ : Union[str, Any]=0.0_2 ,lowercase__ : Any=3 ,lowercase__ : Tuple=4 ,lowercase__ : Dict=None ,lowercase__ : List[Any]=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Optional[int] ,lowercase__ : List[str] ,lowercase__ : str ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,lowercase__ : Union[str, Any] ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : Optional[int] ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : List[Any] ,lowercase__ : int ,lowercase__ : Tuple ,lowercase__ : List[str] ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Tuple ,lowercase__ : Tuple ,lowercase__ : Tuple ,lowercase__ : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Any ,lowercase__ : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : str ,lowercase__ : Any ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : str ,lowercase__ : Any ,lowercase__ : int ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : int = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Tuple = False def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
52
0
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class a__( lowerCamelCase__ ): def __init__( self : Optional[Any] , __snake_case : List[Any] ): a : Dict = data def __iter__( self : Tuple ): for element in self.data: yield element def lowerCamelCase__ ( _A=True ): a : str = Accelerator(even_batches=_A ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def lowerCamelCase__ ( _A , _A , _A , _A = False ): if iterable: a : Dict = DummyIterableDataset(torch.as_tensor(range(_A ) ) ) else: a : Any = TensorDataset(torch.as_tensor(range(_A ) ) ) a : Any = DataLoader(_A , batch_size=_A ) a : List[str] = accelerator.prepare(_A ) return dl def lowerCamelCase__ ( _A , _A , _A , _A , _A , ): a : Dict = create_dataloader(accelerator=_A , dataset_size=_A , batch_size=_A ) a : Dict = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def lowerCamelCase__ ( ): a : List[Any] = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( _A , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( _A , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def lowerCamelCase__ ( ): a : Dict = create_accelerator(even_batches=_A ) verify_dataloader_batch_sizes( _A , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( _A , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def lowerCamelCase__ ( ): a : Optional[int] = create_accelerator(even_batches=_A ) a : Any = torch.nn.Linear(1 , 1 ) a : Optional[Any] = accelerator.prepare(_A ) a : Dict = create_dataloader(_A , dataset_size=3 , batch_size=1 ) a : List[str] = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(_A ): a : List[Any] = ddp_model(batch[0].float() ) a : Tuple = output.sum() loss.backward() batch_idxs.append(_A ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def lowerCamelCase__ ( _A ): with warnings.catch_warnings(record=_A ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , _A ) assert "only supported for multi-GPU" in str(w[-1].message ) def lowerCamelCase__ ( ): a : Optional[int] = True a : Tuple = False a : Optional[int] = create_accelerator(even_batches=_A ) a : int = torch.nn.Linear(1 , 1 ) a : Optional[int] = accelerator.prepare(_A ) a : str = create_dataloader(_A , dataset_size=3 , batch_size=1 ) a : Dict = create_dataloader(_A , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=_A ): a : Tuple = train_dl.batch_sampler.even_batches a : Optional[int] = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def lowerCamelCase__ ( ): a : Optional[Any] = True a : Tuple = False a : Tuple = create_accelerator(even_batches=_A ) a : int = torch.nn.Linear(1 , 1 ) a : Tuple = accelerator.prepare(_A ) create_dataloader(_A , dataset_size=3 , batch_size=1 , iterable=_A ) a : str = create_dataloader(_A , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings('ignore' ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_A ): a : int = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def lowerCamelCase__ ( ): a : Union[str, Any] = create_accelerator() a : int = torch.nn.Linear(1 , 1 ) a : str = accelerator.prepare(_A ) create_dataloader(_A , dataset_size=3 , batch_size=1 , iterable=_A ) with warnings.catch_warnings(record=_A ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_A ): pass assert issubclass(w[-1].category , _A ) assert "only supported for map-style datasets" in str(w[-1].message ) def lowerCamelCase__ ( ): a : Optional[Any] = create_accelerator() accelerator.print('Test that even_batches variable ensures uniform batches across processes' ) test_default_ensures_even_batch_sizes() accelerator.print('Run tests with even_batches disabled' ) test_can_disable_even_batches() accelerator.print('Test joining uneven inputs' ) test_can_join_uneven_inputs() accelerator.print('Test overriding even_batches when joining uneven inputs' ) test_join_can_override_even_batches() accelerator.print('Test overriding even_batches for mixed dataloader types' ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print('Test overriding even_batches raises a warning for iterable dataloaders' ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print('Test join with non DDP distributed raises warning' ) a : List[Any] = accelerator.state.distributed_type a : str = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(_A ) a : List[str] = original_state if __name__ == "__main__": main()
297
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase: str = { 'configuration_mgp_str': ['MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MgpstrConfig'], 'processing_mgp_str': ['MgpstrProcessor'], 'tokenization_mgp_str': ['MgpstrTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: Optional[Any] = [ 'MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST', 'MgpstrModel', 'MgpstrPreTrainedModel', 'MgpstrForSceneTextRecognition', ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys lowerCAmelCase: Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
297
1
import argparse import json import subprocess def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = [] _SCREAMING_SNAKE_CASE : Optional[int] = ( f"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _SCREAMING_SNAKE_CASE : Optional[int] = subprocess.run(__lowerCamelCase, shell=__lowerCamelCase, stdout=subprocess.PIPE ) _SCREAMING_SNAKE_CASE : Tuple = output.stdout.decode("utf-8" ) _SCREAMING_SNAKE_CASE : Union[str, Any] = json.loads(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__lowerCamelCase ) # save the result so we can report them on Slack with open("offline_runners.txt", "w" ) as fp: fp.write(json.dumps(__lowerCamelCase ) ) if len(__lowerCamelCase ) > 0: _SCREAMING_SNAKE_CASE : int = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def lowerCamelCase__ (__lowerCamelCase ): return values.split("," ) UpperCamelCase__ =argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) UpperCamelCase__ =parser.parse_args() get_runner_status(args.target_runners, args.token)
358
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class lowerCAmelCase__: '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase=1_3 , __lowerCamelCase=7 , __lowerCamelCase=True , __lowerCamelCase=True , __lowerCamelCase=True , __lowerCamelCase=True , __lowerCamelCase=9_9 , __lowerCamelCase=3_2 , __lowerCamelCase=2 , __lowerCamelCase=4 , __lowerCamelCase=3_7 , __lowerCamelCase="gelu" , __lowerCamelCase=0.1 , __lowerCamelCase=0.1 , __lowerCamelCase=5_1_2 , __lowerCamelCase=1_6 , __lowerCamelCase=2 , __lowerCamelCase=0.02 , __lowerCamelCase=3 , __lowerCamelCase=4 , __lowerCamelCase=None , ) -> Any: _SCREAMING_SNAKE_CASE : str = parent _SCREAMING_SNAKE_CASE : List[Any] = 1_3 _SCREAMING_SNAKE_CASE : List[str] = 7 _SCREAMING_SNAKE_CASE : Dict = True _SCREAMING_SNAKE_CASE : List[str] = True _SCREAMING_SNAKE_CASE : int = True _SCREAMING_SNAKE_CASE : Union[str, Any] = True _SCREAMING_SNAKE_CASE : int = 9_9 _SCREAMING_SNAKE_CASE : str = 3_8_4 _SCREAMING_SNAKE_CASE : List[Any] = 2 _SCREAMING_SNAKE_CASE : Dict = 4 _SCREAMING_SNAKE_CASE : Dict = 3_7 _SCREAMING_SNAKE_CASE : Union[str, Any] = "gelu" _SCREAMING_SNAKE_CASE : str = 0.1 _SCREAMING_SNAKE_CASE : str = 0.1 _SCREAMING_SNAKE_CASE : List[Any] = 5_1_2 _SCREAMING_SNAKE_CASE : Tuple = 1_6 _SCREAMING_SNAKE_CASE : Dict = 2 _SCREAMING_SNAKE_CASE : Any = 0.02 _SCREAMING_SNAKE_CASE : Any = 3 _SCREAMING_SNAKE_CASE : List[str] = 4 _SCREAMING_SNAKE_CASE : List[Any] = 1_2_8 _SCREAMING_SNAKE_CASE : Optional[int] = 2 _SCREAMING_SNAKE_CASE : int = 9 _SCREAMING_SNAKE_CASE : List[str] = 1 _SCREAMING_SNAKE_CASE : List[Any] = None def UpperCamelCase_ ( self ) -> List[Any]: _SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE : List[str] = None if self.use_input_mask: _SCREAMING_SNAKE_CASE : Dict = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE : Dict = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE : List[Any] = None _SCREAMING_SNAKE_CASE : Union[str, Any] = None _SCREAMING_SNAKE_CASE : Optional[int] = None if self.use_labels: _SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE : Union[str, Any] = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__lowerCamelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> str: _SCREAMING_SNAKE_CASE : Any = TFConvBertModel(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _SCREAMING_SNAKE_CASE : str = [input_ids, input_mask] _SCREAMING_SNAKE_CASE : Any = model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Optional[int]: _SCREAMING_SNAKE_CASE : Dict = TFConvBertForMaskedLM(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _SCREAMING_SNAKE_CASE : List[str] = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE : int = self.num_labels _SCREAMING_SNAKE_CASE : str = TFConvBertForSequenceClassification(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _SCREAMING_SNAKE_CASE : Any = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> str: _SCREAMING_SNAKE_CASE : Optional[int] = self.num_choices _SCREAMING_SNAKE_CASE : List[Any] = TFConvBertForMultipleChoice(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = tf.tile(tf.expand_dims(__lowerCamelCase , 1 ) , (1, self.num_choices, 1) ) _SCREAMING_SNAKE_CASE : Union[str, Any] = tf.tile(tf.expand_dims(__lowerCamelCase , 1 ) , (1, self.num_choices, 1) ) _SCREAMING_SNAKE_CASE : List[Any] = tf.tile(tf.expand_dims(__lowerCamelCase , 1 ) , (1, self.num_choices, 1) ) _SCREAMING_SNAKE_CASE : List[Any] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _SCREAMING_SNAKE_CASE : List[Any] = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Optional[int]: _SCREAMING_SNAKE_CASE : Dict = self.num_labels _SCREAMING_SNAKE_CASE : Tuple = TFConvBertForTokenClassification(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _SCREAMING_SNAKE_CASE : int = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> int: _SCREAMING_SNAKE_CASE : Optional[int] = TFConvBertForQuestionAnswering(config=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _SCREAMING_SNAKE_CASE : Any = model(__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self ) -> Tuple: _SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) : List[Any] = config_and_inputs _SCREAMING_SNAKE_CASE : Optional[Any] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class lowerCAmelCase__( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' __snake_case = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __snake_case = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __snake_case = False __snake_case = False __snake_case = False def UpperCamelCase_ ( self ) -> str: _SCREAMING_SNAKE_CASE : int = TFConvBertModelTester(self ) _SCREAMING_SNAKE_CASE : int = ConfigTester(self , config_class=__lowerCamelCase , hidden_size=3_7 ) def UpperCamelCase_ ( self ) -> List[Any]: self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def UpperCamelCase_ ( self ) -> Dict: _SCREAMING_SNAKE_CASE : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowerCamelCase ) def UpperCamelCase_ ( self ) -> List[Any]: _SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__lowerCamelCase ) def UpperCamelCase_ ( self ) -> Dict: _SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowerCamelCase ) def UpperCamelCase_ ( self ) -> Optional[Any]: _SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowerCamelCase ) def UpperCamelCase_ ( self ) -> int: _SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowerCamelCase ) @slow def UpperCamelCase_ ( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE : Union[str, Any] = True _SCREAMING_SNAKE_CASE : Any = True if hasattr(__lowerCamelCase , "use_cache" ): _SCREAMING_SNAKE_CASE : List[str] = True _SCREAMING_SNAKE_CASE : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _SCREAMING_SNAKE_CASE : Any = getattr(self.model_tester , "key_length" , __lowerCamelCase ) for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE : Union[str, Any] = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = model_class(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Union[str, Any] = len(model(__lowerCamelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowerCamelCase , saved_model=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = os.path.join(__lowerCamelCase , "saved_model" , "1" ) _SCREAMING_SNAKE_CASE : Optional[Any] = tf.keras.models.load_model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = model(__lowerCamelCase ) if self.is_encoder_decoder: _SCREAMING_SNAKE_CASE : List[Any] = outputs["encoder_hidden_states"] _SCREAMING_SNAKE_CASE : Union[str, Any] = outputs["encoder_attentions"] else: _SCREAMING_SNAKE_CASE : List[str] = outputs["hidden_states"] _SCREAMING_SNAKE_CASE : Dict = outputs["attentions"] self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(__lowerCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def UpperCamelCase_ ( self ) -> str: _SCREAMING_SNAKE_CASE : Any = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(__lowerCamelCase ) def UpperCamelCase_ ( self ) -> Dict: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE : Dict = True _SCREAMING_SNAKE_CASE : Dict = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) _SCREAMING_SNAKE_CASE : Dict = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _SCREAMING_SNAKE_CASE : Any = getattr(self.model_tester , "key_length" , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Union[str, Any] = getattr(self.model_tester , "key_length" , __lowerCamelCase ) def check_decoder_attentions_output(__lowerCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = len(__lowerCamelCase ) self.assertEqual(out_len % 2 , 0 ) _SCREAMING_SNAKE_CASE : Optional[int] = outputs.decoder_attentions self.assertEqual(len(__lowerCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(__lowerCamelCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__lowerCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE : Any = True _SCREAMING_SNAKE_CASE : Any = False _SCREAMING_SNAKE_CASE : Optional[Any] = model_class(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = model(self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) _SCREAMING_SNAKE_CASE : Any = len(__lowerCamelCase ) self.assertEqual(config.output_hidden_states , __lowerCamelCase ) check_encoder_attentions_output(__lowerCamelCase ) if self.is_encoder_decoder: _SCREAMING_SNAKE_CASE : Tuple = model_class(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = model(self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) self.assertEqual(config.output_hidden_states , __lowerCamelCase ) check_decoder_attentions_output(__lowerCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _SCREAMING_SNAKE_CASE : Dict = True _SCREAMING_SNAKE_CASE : List[Any] = model_class(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = model(self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) self.assertEqual(config.output_hidden_states , __lowerCamelCase ) check_encoder_attentions_output(__lowerCamelCase ) # Check attention is always last and order is fine _SCREAMING_SNAKE_CASE : Union[str, Any] = True _SCREAMING_SNAKE_CASE : Any = True _SCREAMING_SNAKE_CASE : Optional[int] = model_class(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = model(self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__lowerCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __lowerCamelCase ) check_encoder_attentions_output(__lowerCamelCase ) @require_tf class lowerCAmelCase__( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self ) -> List[Any]: _SCREAMING_SNAKE_CASE : int = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) _SCREAMING_SNAKE_CASE : Tuple = tf.constant([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE : str = model(__lowerCamelCase )[0] _SCREAMING_SNAKE_CASE : int = [1, 6, 7_6_8] self.assertEqual(output.shape , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = tf.constant( [ [ [-0.0347_5493, -0.468_6034, -0.3063_8832], [0.2263_7248, -0.2698_8646, -0.742_3424], [0.1032_4868, -0.4501_3508, -0.5828_0784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __lowerCamelCase , atol=1E-4 )
325
0
'''simple docstring''' from __future__ import annotations def a__ ( lowercase : float, lowercase : float, lowercase : float, ) -> tuple[str, float]: """simple docstring""" if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif stress < 0: raise ValueError('''Stress cannot be negative''' ) elif tangential_force < 0: raise ValueError('''Tangential Force cannot be negative''' ) elif area < 0: raise ValueError('''Area cannot be negative''' ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
324
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : List[str] = logging.get_logger(__name__) lowercase__ : Optional[int] = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : int = 'audio-spectrogram-transformer' def __init__( self : Optional[Any] , lowerCAmelCase__ : List[str]=768 , lowerCAmelCase__ : Optional[Any]=12 , lowerCAmelCase__ : int=12 , lowerCAmelCase__ : int=3072 , lowerCAmelCase__ : List[str]="gelu" , lowerCAmelCase__ : List[Any]=0.0 , lowerCAmelCase__ : Optional[Any]=0.0 , lowerCAmelCase__ : int=0.02 , lowerCAmelCase__ : Union[str, Any]=1e-1_2 , lowerCAmelCase__ : Any=16 , lowerCAmelCase__ : str=True , lowerCAmelCase__ : List[str]=10 , lowerCAmelCase__ : int=10 , lowerCAmelCase__ : Dict=1024 , lowerCAmelCase__ : Optional[int]=128 , **lowerCAmelCase__ : List[Any] , ) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = patch_size _UpperCamelCase = qkv_bias _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins
324
1
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = [ ['''attention''', '''attn'''], ['''encoder_attention''', '''encoder_attn'''], ['''q_lin''', '''q_proj'''], ['''k_lin''', '''k_proj'''], ['''v_lin''', '''v_proj'''], ['''out_lin''', '''out_proj'''], ['''norm_embeddings''', '''layernorm_embedding'''], ['''position_embeddings''', '''embed_positions'''], ['''embeddings''', '''embed_tokens'''], ['''ffn.lin''', '''fc'''], ] def _snake_case ( A ) -> Union[str, Any]: if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: lowerCAmelCase__ = k.replace(A , A ) if k.startswith('''encoder''' ): lowerCAmelCase__ = k.replace('''.attn''' , '''.self_attn''' ) lowerCAmelCase__ = k.replace('''norm1''' , '''self_attn_layer_norm''' ) lowerCAmelCase__ = k.replace('''norm2''' , '''final_layer_norm''' ) elif k.startswith('''decoder''' ): lowerCAmelCase__ = k.replace('''norm1''' , '''self_attn_layer_norm''' ) lowerCAmelCase__ = k.replace('''norm2''' , '''encoder_attn_layer_norm''' ) lowerCAmelCase__ = k.replace('''norm3''' , '''final_layer_norm''' ) return k def _snake_case ( A ) -> List[Any]: lowerCAmelCase__ = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: lowerCAmelCase__ = sd.pop(A ) lowerCAmelCase__ = k.replace('''layernorm_embedding''' , '''layer_norm''' ) assert new_k not in sd lowerCAmelCase__ = v __UpperCAmelCase = ['''START'''] @torch.no_grad() def _snake_case ( A , A , A ) -> List[Any]: lowerCAmelCase__ = torch.load(A , map_location='''cpu''' ) lowerCAmelCase__ = model['''model'''] lowerCAmelCase__ = BlenderbotConfig.from_json_file(A ) lowerCAmelCase__ = BlenderbotForConditionalGeneration(A ) lowerCAmelCase__ = m.model.state_dict().keys() lowerCAmelCase__ = [] lowerCAmelCase__ = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue lowerCAmelCase__ = rename_state_dict_key(A ) if new_k not in valid_keys: failures.append([k, new_k] ) else: lowerCAmelCase__ = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(A ) m.model.load_state_dict(A , strict=A ) m.half() m.save_pretrained(A ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''--src_path''', type=str, help='''like blenderbot-model.bin''') parser.add_argument('''--save_dir''', default='''hf_blenderbot''', type=str, help='''Where to save converted model.''') parser.add_argument( '''--hf_config_json''', default='''blenderbot-3b-config.json''', type=str, help='''Path to config to use''' ) __UpperCAmelCase = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
228
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __UpperCAmelCase = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' __UpperCAmelCase = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' __UpperCAmelCase = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def _snake_case ( A , A ) -> List[Any]: return float((preds == labels).mean() ) def _snake_case ( A , A , A="binary" ) -> int: lowerCAmelCase__ = simple_accuracy(A , A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=A , average=A ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( A , A ) -> List[Any]: lowerCAmelCase__ = {} for id_pred, label in zip(A , A ): lowerCAmelCase__ = F"""{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}""" lowerCAmelCase__ = id_pred['''prediction'''] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase__ = [(pred, label)] lowerCAmelCase__ , lowerCAmelCase__ = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase__ , lowerCAmelCase__ = zip(*A ) lowerCAmelCase__ = fa_score(y_true=A , y_pred=A , average='''macro''' ) fas.append(A ) lowerCAmelCase__ = int(sum(pred == label for pred, label in preds_labels ) == len(A ) ) ems.append(A ) lowerCAmelCase__ = float(sum(A ) / len(A ) ) lowerCAmelCase__ = sum(A ) / len(A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=[id_pred['''prediction'''] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self ) -> Dict: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if not self.config_name == '''record''' and not self.config_name == '''multirc''' else None , ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "prediction_text": datasets.Value('''string''' ), }, "references": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "answers": datasets.Sequence(datasets.Value('''string''' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('''int64''' ), "paragraph": datasets.Value('''int64''' ), "question": datasets.Value('''int64''' ), }, "prediction": datasets.Value('''int64''' ), }, "references": datasets.Value('''int64''' ), } else: return { "predictions": datasets.Value('''int64''' ), "references": datasets.Value('''int64''' ), } def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCamelCase_ , lowerCamelCase_ )} elif self.config_name == "cb": return acc_and_fa(lowerCamelCase_ , lowerCamelCase_ , fa_avg='''macro''' ) elif self.config_name == "record": lowerCAmelCase__ = [ { '''qas''': [ {'''id''': ref['''idx''']['''query'''], '''answers''': [{'''text''': ans} for ans in ref['''answers''']]} for ref in references ] } ] lowerCAmelCase__ = {pred['''idx''']['''query''']: pred['''prediction_text'''] for pred in predictions} return evaluate_record(lowerCamelCase_ , lowerCamelCase_ )[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCamelCase_ , lowerCamelCase_ ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCamelCase_ , lowerCamelCase_ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' )
228
1
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class snake_case_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : List[str] , _UpperCamelCase : Tuple , _UpperCamelCase : Any=7 , _UpperCamelCase : Tuple=3 , _UpperCamelCase : Dict=1_8 , _UpperCamelCase : str=3_0 , _UpperCamelCase : str=4_0_0 , _UpperCamelCase : Union[str, Any]=True , _UpperCamelCase : List[str]=None , _UpperCamelCase : Tuple=True , ) ->Dict: snake_case_ = size if size is not None else {'''height''': 1_8, '''width''': 1_8} snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = image_size snake_case_ = min_resolution snake_case_ = max_resolution snake_case_ = do_resize snake_case_ = size snake_case_ = apply_ocr def snake_case__( self : str ) ->str: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class snake_case_ ( __A , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = LayoutLMvaImageProcessor if is_pytesseract_available() else None def snake_case__( self : Optional[Any] ) ->List[str]: snake_case_ = LayoutLMvaImageProcessingTester(self ) @property def snake_case__( self : Optional[int] ) ->Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def snake_case__( self : Optional[Any] ) ->str: snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCamelCase , '''do_resize''' ) ) self.assertTrue(hasattr(_UpperCamelCase , '''size''' ) ) self.assertTrue(hasattr(_UpperCamelCase , '''apply_ocr''' ) ) def snake_case__( self : Optional[int] ) ->Tuple: snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 1_8, '''width''': 1_8} ) snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {'''height''': 4_2, '''width''': 4_2} ) def snake_case__( self : List[str] ) ->List[str]: pass def snake_case__( self : str ) ->Tuple: snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , Image.Image ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , _UpperCamelCase ) self.assertIsInstance(encoding.boxes , _UpperCamelCase ) # Test batched snake_case_ = image_processing(_UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def snake_case__( self : List[Any] ) ->List[str]: snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase , numpify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , np.ndarray ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case_ = image_processing(_UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def snake_case__( self : str ) ->List[str]: snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase , torchify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , torch.Tensor ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case_ = image_processing(_UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def snake_case__( self : Optional[int] ) ->Optional[int]: snake_case_ = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case_ = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case_ = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case_ = image_processing(_UpperCamelCase , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case_ = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case_ = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , _UpperCamelCase ) self.assertListEqual(encoding.boxes , _UpperCamelCase ) # with apply_OCR = False snake_case_ = LayoutLMvaImageProcessor(apply_ocr=_UpperCamelCase ) snake_case_ = image_processing(_UpperCamelCase , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_2_4, 2_2_4) )
8
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class __A ( A ): '''simple docstring''' __lowerCamelCase : Optional[Any] = 'MCTCTFeatureExtractor' __lowerCamelCase : Optional[Any] = 'AutoTokenizer' def __init__(self , A , A ) -> Dict: """simple docstring""" super().__init__(A , A ) _a = self.feature_extractor _a = False def __call__(self , *A , **A ) -> Optional[int]: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*A , **A ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) _a = kwargs.pop('''raw_speech''' ) else: _a = kwargs.pop('''audio''' , A ) _a = kwargs.pop('''sampling_rate''' , A ) _a = kwargs.pop('''text''' , A ) if len(A ) > 0: _a = args[0] _a = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: _a = self.feature_extractor(A , *A , sampling_rate=A , **A ) if text is not None: _a = self.tokenizer(A , **A ) if text is None: return inputs elif audio is None: return encodings else: _a = encodings['''input_ids'''] return inputs def a__ (self , *A , **A ) -> int: """simple docstring""" return self.tokenizer.batch_decode(*A , **A ) def a__ (self , *A , **A ) -> Tuple: """simple docstring""" if self._in_target_context_manager: return self.current_processor.pad(*A , **A ) _a = kwargs.pop('''input_features''' , A ) _a = kwargs.pop('''labels''' , A ) if len(A ) > 0: _a = args[0] _a = args[1:] if input_features is not None: _a = self.feature_extractor.pad(A , *A , **A ) if labels is not None: _a = self.tokenizer.pad(A , **A ) if labels is None: return input_features elif input_features is None: return labels else: _a = labels['''input_ids'''] return input_features def a__ (self , *A , **A ) -> List[str]: """simple docstring""" return self.tokenizer.decode(*A , **A ) @contextmanager def a__ (self ) -> Dict: """simple docstring""" warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) _a = True _a = self.tokenizer yield _a = self.feature_extractor _a = False
211
0
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __UpperCAmelCase ( lowerCamelCase__ ): UpperCamelCase = """ClapFeatureExtractor""" UpperCamelCase = ("""RobertaTokenizer""", """RobertaTokenizerFast""") def __init__( self : str, __A : Optional[Any], __A : str ): super().__init__(__A, __A ) def __call__( self : Any, __A : int=None, __A : List[str]=None, __A : Union[str, Any]=None, **__A : str ): UpperCAmelCase : Tuple = kwargs.pop('''sampling_rate''', __A ) if text is None and audios is None: raise ValueError('''You have to specify either text or audios. Both cannot be none.''' ) if text is not None: UpperCAmelCase : Optional[Any] = self.tokenizer(__A, return_tensors=__A, **__A ) if audios is not None: UpperCAmelCase : Any = self.feature_extractor( __A, sampling_rate=__A, return_tensors=__A, **__A ) if text is not None and audios is not None: UpperCAmelCase : int = audio_features.input_features return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__A ), tensor_type=__A ) def __magic_name__ ( self : int, *__A : List[Any], **__A : Union[str, Any] ): return self.tokenizer.batch_decode(*__A, **__A ) def __magic_name__ ( self : List[str], *__A : int, **__A : Any ): return self.tokenizer.decode(*__A, **__A ) @property def __magic_name__ ( self : Optional[Any] ): UpperCAmelCase : Optional[int] = self.tokenizer.model_input_names UpperCAmelCase : List[str] = self.feature_extractor.model_input_names return list(dict.fromkeys(tokenizer_input_names + feature_extractor_input_names ) )
99
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCAmelCase ( lowerCamelCase__ ): def __init__( self : Dict, __A : Any, __A : Optional[int]=1_3, __A : Any=7, __A : Tuple=True, __A : int=True, __A : Dict=True, __A : Union[str, Any]=True, __A : Optional[int]=9_9, __A : Optional[int]=3_2, __A : Union[str, Any]=5, __A : Optional[int]=4, __A : str=3_7, __A : Union[str, Any]="gelu", __A : Optional[int]=0.1, __A : Optional[Any]=0.1, __A : Any=5_1_2, __A : List[str]=1_6, __A : Optional[int]=2, __A : Union[str, Any]=0.0_2, __A : Optional[int]=False, __A : List[str]=True, __A : int="None", __A : List[str]=3, __A : Any=4, __A : Dict=None, ): UpperCAmelCase : str = parent UpperCAmelCase : int = batch_size UpperCAmelCase : Tuple = seq_length UpperCAmelCase : Union[str, Any] = is_training UpperCAmelCase : Dict = use_input_mask UpperCAmelCase : Optional[Any] = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : List[Any] = vocab_size UpperCAmelCase : Union[str, Any] = hidden_size UpperCAmelCase : Tuple = num_hidden_layers UpperCAmelCase : str = num_attention_heads UpperCAmelCase : Tuple = intermediate_size UpperCAmelCase : Optional[Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : int = type_vocab_size UpperCAmelCase : str = type_sequence_label_size UpperCAmelCase : Union[str, Any] = initializer_range UpperCAmelCase : Dict = num_labels UpperCAmelCase : Optional[Any] = num_choices UpperCAmelCase : str = relative_attention UpperCAmelCase : Any = position_biased_input UpperCAmelCase : str = pos_att_type UpperCAmelCase : Union[str, Any] = scope def __magic_name__ ( self : Union[str, Any] ): UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCAmelCase : int = None if self.use_input_mask: UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length], vocab_size=2 ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size ) UpperCAmelCase : List[str] = None UpperCAmelCase : str = None UpperCAmelCase : Any = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size], self.type_sequence_label_size ) UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) UpperCAmelCase : List[Any] = ids_tensor([self.batch_size], self.num_choices ) UpperCAmelCase : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __magic_name__ ( self : Any ): return DebertaVaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, relative_attention=self.relative_attention, position_biased_input=self.position_biased_input, pos_att_type=self.pos_att_type, ) def __magic_name__ ( self : Dict, __A : str ): self.parent.assertListEqual(list(result.loss.size() ), [] ) def __magic_name__ ( self : List[str], __A : Dict, __A : int, __A : str, __A : List[str], __A : Dict, __A : str, __A : int ): UpperCAmelCase : Optional[int] = DebertaVaModel(config=__A ) model.to(__A ) model.eval() UpperCAmelCase : Optional[int] = model(__A, attention_mask=__A, token_type_ids=__A )[0] UpperCAmelCase : Optional[int] = model(__A, token_type_ids=__A )[0] UpperCAmelCase : int = model(__A )[0] self.parent.assertListEqual(list(sequence_output.size() ), [self.batch_size, self.seq_length, self.hidden_size] ) def __magic_name__ ( self : Dict, __A : Union[str, Any], __A : Optional[Any], __A : Tuple, __A : Optional[int], __A : List[Any], __A : List[Any], __A : Optional[int] ): UpperCAmelCase : int = DebertaVaForMaskedLM(config=__A ) model.to(__A ) model.eval() UpperCAmelCase : int = model(__A, attention_mask=__A, token_type_ids=__A, labels=__A ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def __magic_name__ ( self : List[str], __A : str, __A : Optional[Any], __A : List[str], __A : Optional[int], __A : List[Any], __A : int, __A : Optional[int] ): UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = DebertaVaForSequenceClassification(__A ) model.to(__A ) model.eval() UpperCAmelCase : int = model(__A, attention_mask=__A, token_type_ids=__A, labels=__A ) self.parent.assertListEqual(list(result.logits.size() ), [self.batch_size, self.num_labels] ) self.check_loss_output(__A ) def __magic_name__ ( self : Any, __A : Tuple, __A : Any, __A : str, __A : List[Any], __A : Dict, __A : Optional[Any], __A : List[str] ): UpperCAmelCase : Dict = self.num_labels UpperCAmelCase : int = DebertaVaForTokenClassification(config=__A ) model.to(__A ) model.eval() UpperCAmelCase : Tuple = model(__A, attention_mask=__A, token_type_ids=__A, labels=__A ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def __magic_name__ ( self : Tuple, __A : List[str], __A : Tuple, __A : Tuple, __A : int, __A : Optional[Any], __A : Tuple, __A : Any ): UpperCAmelCase : Union[str, Any] = DebertaVaForQuestionAnswering(config=__A ) model.to(__A ) model.eval() UpperCAmelCase : Any = model( __A, attention_mask=__A, token_type_ids=__A, start_positions=__A, end_positions=__A, ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def __magic_name__ ( self : Dict, __A : Optional[int], __A : str, __A : List[str], __A : Dict, __A : Optional[Any], __A : Union[str, Any], __A : int ): UpperCAmelCase : Union[str, Any] = DebertaVaForMultipleChoice(config=__A ) model.to(__A ) model.eval() UpperCAmelCase : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() UpperCAmelCase : int = token_type_ids.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() UpperCAmelCase : Tuple = input_mask.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() UpperCAmelCase : int = model( __A, attention_mask=__A, token_type_ids=__A, labels=__A, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices) ) def __magic_name__ ( self : List[str] ): UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : int = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) UpperCamelCase = ( { """feature-extraction""": DebertaVaModel, """fill-mask""": DebertaVaForMaskedLM, """question-answering""": DebertaVaForQuestionAnswering, """text-classification""": DebertaVaForSequenceClassification, """token-classification""": DebertaVaForTokenClassification, """zero-shot""": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase = True UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def __magic_name__ ( self : Union[str, Any] ): UpperCAmelCase : str = DebertaVaModelTester(self ) UpperCAmelCase : Dict = ConfigTester(self, config_class=__A, hidden_size=3_7 ) def __magic_name__ ( self : Any ): self.config_tester.run_common_tests() def __magic_name__ ( self : Union[str, Any] ): UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__A ) def __magic_name__ ( self : List[Any] ): UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__A ) def __magic_name__ ( self : Tuple ): UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__A ) def __magic_name__ ( self : Union[str, Any] ): UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__A ) def __magic_name__ ( self : List[str] ): UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__A ) def __magic_name__ ( self : Any ): UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*__A ) @slow def __magic_name__ ( self : Dict ): for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = DebertaVaModel.from_pretrained(__A ) self.assertIsNotNone(__A ) @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def __magic_name__ ( self : str ): pass @slow def __magic_name__ ( self : Optional[int] ): UpperCAmelCase : str = DebertaVaModel.from_pretrained('''microsoft/deberta-v2-xlarge''' ) UpperCAmelCase : Union[str, Any] = torch.tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCAmelCase : List[str] = model(__A, attention_mask=__A )[0] # compare the actual values for a slice. UpperCAmelCase : List[str] = torch.tensor( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4], __A, atol=1E-4 ), F'''{output[:, 1:4, 1:4]}''' )
99
1
"""simple docstring""" import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) def lowerCamelCase ( _UpperCamelCase : Dict ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : str = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: __UpperCAmelCase : List[Any] = 1_2_8 elif "12-12" in model_name: __UpperCAmelCase : Optional[Any] = 1_2 __UpperCAmelCase : Dict = 1_2 elif "14-14" in model_name: __UpperCAmelCase : str = 1_4 __UpperCAmelCase : Any = 1_4 elif "16-16" in model_name: __UpperCAmelCase : Optional[Any] = 1_6 __UpperCAmelCase : Union[str, Any] = 1_6 else: raise ValueError("""Model not supported""" ) __UpperCAmelCase : Any = """huggingface/label-files""" if "speech-commands" in model_name: __UpperCAmelCase : Dict = 3_5 __UpperCAmelCase : Union[str, Any] = """speech-commands-v2-id2label.json""" else: __UpperCAmelCase : Optional[int] = 5_2_7 __UpperCAmelCase : List[Any] = """audioset-id2label.json""" __UpperCAmelCase : List[str] = json.load(open(hf_hub_download(_UpperCamelCase , _UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) __UpperCAmelCase : List[str] = {int(_UpperCamelCase ): v for k, v in idalabel.items()} __UpperCAmelCase : List[Any] = idalabel __UpperCAmelCase : Optional[Any] = {v: k for k, v in idalabel.items()} return config def lowerCamelCase ( _UpperCamelCase : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if "module.v" in name: __UpperCAmelCase : Optional[int] = name.replace("""module.v""" , """audio_spectrogram_transformer""" ) if "cls_token" in name: __UpperCAmelCase : Tuple = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "dist_token" in name: __UpperCAmelCase : Optional[int] = name.replace("""dist_token""" , """embeddings.distillation_token""" ) if "pos_embed" in name: __UpperCAmelCase : Union[str, Any] = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: __UpperCAmelCase : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) # transformer blocks if "blocks" in name: __UpperCAmelCase : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: __UpperCAmelCase : Optional[int] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: __UpperCAmelCase : List[str] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: __UpperCAmelCase : Optional[Any] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __UpperCAmelCase : Union[str, Any] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: __UpperCAmelCase : Any = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __UpperCAmelCase : Dict = name.replace("""mlp.fc2""" , """output.dense""" ) # final layernorm if "audio_spectrogram_transformer.norm" in name: __UpperCAmelCase : List[Any] = name.replace("""audio_spectrogram_transformer.norm""" , """audio_spectrogram_transformer.layernorm""" ) # classifier head if "module.mlp_head.0" in name: __UpperCAmelCase : Union[str, Any] = name.replace("""module.mlp_head.0""" , """classifier.layernorm""" ) if "module.mlp_head.1" in name: __UpperCAmelCase : Union[str, Any] = name.replace("""module.mlp_head.1""" , """classifier.dense""" ) return name def lowerCamelCase ( _UpperCamelCase : Any , _UpperCamelCase : Dict ) -> Any: '''simple docstring''' for key in orig_state_dict.copy().keys(): __UpperCAmelCase : Any = orig_state_dict.pop(_UpperCamelCase ) if "qkv" in key: __UpperCAmelCase : int = key.split(""".""" ) __UpperCAmelCase : List[str] = int(key_split[3] ) __UpperCAmelCase : Any = config.hidden_size if "weight" in key: __UpperCAmelCase : List[Any] = val[:dim, :] __UpperCAmelCase : Optional[int] = val[dim : dim * 2, :] __UpperCAmelCase : Dict = val[-dim:, :] else: __UpperCAmelCase : Optional[Any] = val[:dim] __UpperCAmelCase : Any = val[dim : dim * 2] __UpperCAmelCase : Union[str, Any] = val[-dim:] else: __UpperCAmelCase : Optional[Any] = val return orig_state_dict def lowerCamelCase ( _UpperCamelCase : Optional[Any] ) -> Any: '''simple docstring''' __UpperCAmelCase : Dict = [ """module.v.head.weight""", """module.v.head.bias""", """module.v.head_dist.weight""", """module.v.head_dist.bias""", ] for k in ignore_keys: state_dict.pop(_UpperCamelCase , _UpperCamelCase ) @torch.no_grad() def lowerCamelCase ( _UpperCamelCase : Dict , _UpperCamelCase : List[Any] , _UpperCamelCase : Any=False ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : Optional[int] = get_audio_spectrogram_transformer_config(_UpperCamelCase ) __UpperCAmelCase : Any = { """ast-finetuned-audioset-10-10-0.4593""": ( """https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.450""": ( """https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.448""": ( """https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.448-v2""": ( """https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1""" ), """ast-finetuned-audioset-12-12-0.447""": ( """https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1""" ), """ast-finetuned-audioset-14-14-0.443""": ( """https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1""" ), """ast-finetuned-audioset-16-16-0.442""": ( """https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1""" ), """ast-finetuned-speech-commands-v2""": ( """https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1""" ), } # load original state_dict __UpperCAmelCase : Dict = model_name_to_url[model_name] __UpperCAmelCase : Any = torch.hub.load_state_dict_from_url(_UpperCamelCase , map_location="""cpu""" ) # remove some keys remove_keys(_UpperCamelCase ) # rename some keys __UpperCAmelCase : Union[str, Any] = convert_state_dict(_UpperCamelCase , _UpperCamelCase ) # load 🤗 model __UpperCAmelCase : Tuple = ASTForAudioClassification(_UpperCamelCase ) model.eval() model.load_state_dict(_UpperCamelCase ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 __UpperCAmelCase : Optional[Any] = -4.2_677_393 if """speech-commands""" not in model_name else -6.845_978 __UpperCAmelCase : List[Any] = 4.5_689_974 if """speech-commands""" not in model_name else 5.5_654_526 __UpperCAmelCase : List[Any] = 1_0_2_4 if """speech-commands""" not in model_name else 1_2_8 __UpperCAmelCase : str = ASTFeatureExtractor(mean=_UpperCamelCase , std=_UpperCamelCase , max_length=_UpperCamelCase ) if "speech-commands" in model_name: __UpperCAmelCase : Tuple = load_dataset("""speech_commands""" , """v0.02""" , split="""validation""" ) __UpperCAmelCase : Optional[int] = dataset[0]["""audio"""]["""array"""] else: __UpperCAmelCase : Any = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" , ) __UpperCAmelCase ,__UpperCAmelCase : Tuple = torchaudio.load(_UpperCamelCase ) __UpperCAmelCase : Optional[int] = waveform.squeeze().numpy() __UpperCAmelCase : Optional[Any] = feature_extractor(_UpperCamelCase , sampling_rate=1_6_0_0_0 , return_tensors="""pt""" ) # forward pass __UpperCAmelCase : int = model(**_UpperCamelCase ) __UpperCAmelCase : Any = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": __UpperCAmelCase : int = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": __UpperCAmelCase : Tuple = torch.tensor([-1.1_986, -7.0_903, -8.2_718] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": __UpperCAmelCase : Any = torch.tensor([-2.6_128, -8.0_080, -9.4_344] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": __UpperCAmelCase : List[Any] = torch.tensor([-1.5_080, -7.4_534, -8.8_917] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": __UpperCAmelCase : str = torch.tensor([-0.5_050, -6.5_833, -8.0_843] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": __UpperCAmelCase : List[str] = torch.tensor([-0.3_826, -7.0_336, -8.2_413] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": __UpperCAmelCase : int = torch.tensor([-1.2_113, -6.9_101, -8.3_470] ) elif model_name == "ast-finetuned-speech-commands-v2": __UpperCAmelCase : Any = torch.tensor([6.1_589, -8.0_566, -8.7_984] ) else: raise ValueError("""Unknown model name""" ) if not torch.allclose(logits[0, :3] , _UpperCamelCase , atol=1E-4 ): raise ValueError("""Logits don't match""" ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_UpperCamelCase ) print(f'''Saving feature extractor to {pytorch_dump_folder_path}''' ) feature_extractor.save_pretrained(_UpperCamelCase ) if push_to_hub: print("""Pushing model and feature extractor to the hub...""" ) model.push_to_hub(f'''MIT/{model_name}''' ) feature_extractor.push_to_hub(f'''MIT/{model_name}''' ) if __name__ == "__main__": UpperCAmelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='ast-finetuned-audioset-10-10-0.4593', type=str, help='Name of the Audio Spectrogram Transformer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) UpperCAmelCase : Dict = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
115
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class lowerCamelCase__ ( nn.Module ): """simple docstring""" def __init__( self : List[Any] ): '''simple docstring''' super().__init__() __UpperCAmelCase : Optional[Any] = nn.Linear(3 , 4 ) __UpperCAmelCase : Optional[int] = nn.BatchNormad(4 ) __UpperCAmelCase : int = nn.Linear(4 , 5 ) def lowerCamelCase__ ( self : List[str] , UpperCamelCase : Dict ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(UpperCamelCase ) ) ) class lowerCamelCase__ ( A ): """simple docstring""" def lowerCamelCase__ ( self : str , UpperCamelCase : Any , *UpperCamelCase : List[Any] , **UpperCamelCase : Dict ): '''simple docstring''' return (args[0] + 1,) + args[1:], kwargs class lowerCamelCase__ ( A ): """simple docstring""" def lowerCamelCase__ ( self : Any , UpperCamelCase : Dict , UpperCamelCase : List[Any] ): '''simple docstring''' return output + 1 class lowerCamelCase__ ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self : str ): '''simple docstring''' __UpperCAmelCase : Dict = ModelForTest() __UpperCAmelCase : List[str] = ModelHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) self.assertEqual(test_model._hf_hook , UpperCamelCase ) self.assertTrue(hasattr(UpperCamelCase , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCamelCase ) self.assertFalse(hasattr(UpperCamelCase , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCamelCase , """_old_forward""" ) ) def lowerCamelCase__ ( self : str ): '''simple docstring''' __UpperCAmelCase : int = ModelForTest() __UpperCAmelCase : Optional[Any] = ModelHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) add_hook_to_module(UpperCamelCase , UpperCamelCase , append=UpperCamelCase ) self.assertEqual(isinstance(test_model._hf_hook , UpperCamelCase ) , UpperCamelCase ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(UpperCamelCase , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCamelCase ) self.assertFalse(hasattr(UpperCamelCase , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCamelCase , """_old_forward""" ) ) def lowerCamelCase__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = ModelForTest() __UpperCAmelCase : str = torch.randn(2 , 3 ) __UpperCAmelCase : List[str] = test_model(x + 1 ) __UpperCAmelCase : Optional[int] = test_model(x + 2 ) __UpperCAmelCase : Optional[Any] = PreForwardHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = test_model(UpperCamelCase ) self.assertTrue(torch.allclose(UpperCamelCase , UpperCamelCase , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain __UpperCAmelCase : Optional[Any] = PreForwardHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[int] = test_model(UpperCamelCase ) self.assertTrue(torch.allclose(UpperCamelCase , UpperCamelCase , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks __UpperCAmelCase : Optional[Any] = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : List[str] = test_model(UpperCamelCase ) assert torch.allclose(UpperCamelCase , UpperCamelCase , atol=1e-5 ) def lowerCamelCase__ ( self : Dict ): '''simple docstring''' __UpperCAmelCase : Optional[int] = ModelForTest() __UpperCAmelCase : str = torch.randn(2 , 3 ) __UpperCAmelCase : int = test_model(UpperCamelCase ) __UpperCAmelCase : Dict = PostForwardHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Any = test_model(UpperCamelCase ) self.assertTrue(torch.allclose(UpperCamelCase , output + 1 , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain __UpperCAmelCase : List[str] = PostForwardHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = test_model(UpperCamelCase ) self.assertTrue(torch.allclose(UpperCamelCase , output + 1 , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks __UpperCAmelCase : Optional[int] = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : int = test_model(UpperCamelCase ) assert torch.allclose(UpperCamelCase , output + 2 , atol=1e-5 ) def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' __UpperCAmelCase : int = ModelForTest() __UpperCAmelCase : str = torch.randn(2 , 3 ) __UpperCAmelCase : Optional[Any] = test_model(UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = PostForwardHook() add_hook_to_module(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[int] = test_model(UpperCamelCase ) self.assertTrue(torch.allclose(UpperCamelCase , output + 1 ) ) self.assertTrue(outputa.requires_grad ) __UpperCAmelCase : List[str] = True __UpperCAmelCase : Optional[int] = test_model(UpperCamelCase ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def lowerCamelCase__ ( self : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device __UpperCAmelCase : List[str] = torch.randn(2 , 3 ) __UpperCAmelCase : Any = model(UpperCamelCase ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(UpperCamelCase , AlignDevicesHook(io_same_device=UpperCamelCase ) ) __UpperCAmelCase : int = torch.randn(2 , 3 ).to(0 ) __UpperCAmelCase : Any = model(UpperCamelCase ) self.assertEqual(output.device , torch.device(0 ) ) def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' __UpperCAmelCase : Tuple = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __UpperCAmelCase : Union[str, Any] = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCamelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCamelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCamelCase ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __UpperCAmelCase : Any = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = torch.randn(2 , 3 ) __UpperCAmelCase : Optional[int] = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload __UpperCAmelCase : Union[str, Any] = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCamelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCamelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCamelCase ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __UpperCAmelCase : str = torch.randn(2 , 3 ) __UpperCAmelCase : Dict = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def lowerCamelCase__ ( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : int = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __UpperCAmelCase : Optional[int] = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(UpperCamelCase , execution_device=UpperCamelCase , offload=UpperCamelCase ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __UpperCAmelCase : Any = torch.device(UpperCamelCase ) self.assertEqual(model.batchnorm.running_mean.device , UpperCamelCase ) __UpperCAmelCase : str = torch.randn(2 , 3 ) __UpperCAmelCase : Any = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(UpperCamelCase , execution_device=UpperCamelCase , offload=UpperCamelCase , offload_buffers=UpperCamelCase ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __UpperCAmelCase : List[str] = torch.randn(2 , 3 ) __UpperCAmelCase : Any = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def lowerCamelCase__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : Dict = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __UpperCAmelCase : Any = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( UpperCamelCase , execution_device=UpperCamelCase , offload=UpperCamelCase , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __UpperCAmelCase : str = torch.device(UpperCamelCase ) self.assertEqual(model.batchnorm.running_mean.device , UpperCamelCase ) __UpperCAmelCase : Dict = torch.randn(2 , 3 ) __UpperCAmelCase : Optional[Any] = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( UpperCamelCase , execution_device=UpperCamelCase , offload=UpperCamelCase , weights_map=model.state_dict() , offload_buffers=UpperCamelCase , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __UpperCAmelCase : int = torch.randn(2 , 3 ) __UpperCAmelCase : List[str] = model(UpperCamelCase ) self.assertEqual(output.device , UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
115
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=snake_case_ ) class _a ( snake_case_ ): """simple docstring""" _lowerCamelCase : str = field(default='language-modeling' , metadata={'include_in_asdict_even_if_is_default': True} ) _lowerCamelCase : ClassVar[Features] = Features({'text': Value('string' )} ) _lowerCamelCase : ClassVar[Features] = Features({} ) _lowerCamelCase : str = "text" @property def __A ( self : str ): return {self.text_column: "text"}
329
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class _a ( snake_case_ ): """simple docstring""" _lowerCamelCase : torch.FloatTensor _lowerCamelCase : Optional[torch.FloatTensor] = None def __snake_case ( __UpperCamelCase : Tuple ,__UpperCamelCase : Any=0.999 ,__UpperCamelCase : Any="cosine" ,): """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(__UpperCamelCase : Any ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__UpperCamelCase : int ): return math.exp(t * -12.0 ) else: raise ValueError(f'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) A_ = [] for i in range(__UpperCamelCase ): A_ = i / num_diffusion_timesteps A_ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__UpperCamelCase ) / alpha_bar_fn(__UpperCamelCase ) ,__UpperCamelCase ) ) return torch.tensor(__UpperCamelCase ,dtype=torch.floataa ) class _a ( snake_case_ , snake_case_ ): """simple docstring""" @register_to_config def __init__( self : Optional[int] , UpperCAmelCase : int = 1000 , UpperCAmelCase : str = "fixed_small_log" , UpperCAmelCase : bool = True , UpperCAmelCase : Optional[float] = 1.0 , UpperCAmelCase : str = "epsilon" , UpperCAmelCase : str = "squaredcos_cap_v2" , ): if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) A_ = betas_for_alpha_bar(UpperCAmelCase ) A_ = 1.0 - self.betas A_ = torch.cumprod(self.alphas , dim=0 ) A_ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution A_ = 1.0 # setable values A_ = None A_ = torch.from_numpy(np.arange(0 , UpperCAmelCase )[::-1].copy() ) A_ = variance_type def __A ( self : Optional[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Optional[int] = None ): return sample def __A ( self : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Union[str, torch.device] = None ): A_ = num_inference_steps A_ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) A_ = (np.arange(0 , UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) A_ = torch.from_numpy(UpperCAmelCase ).to(UpperCAmelCase ) def __A ( self : List[Any] , UpperCAmelCase : Dict , UpperCAmelCase : str=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=None ): if prev_timestep is None: A_ = t - 1 A_ = self.alphas_cumprod[t] A_ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one A_ = 1 - alpha_prod_t A_ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: A_ = self.betas[t] else: A_ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample A_ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: A_ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": A_ = torch.log(torch.clamp(UpperCAmelCase , min=1E-20 ) ) A_ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler A_ = variance.log() A_ = beta.log() A_ = (predicted_variance + 1) / 2 A_ = frac * max_log + (1 - frac) * min_log return variance def __A ( self : int , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : int , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Dict=None , UpperCAmelCase : bool = True , ): A_ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": A_ , A_ = torch.split(UpperCAmelCase , sample.shape[1] , dim=1 ) else: A_ = None # 1. compute alphas, betas if prev_timestep is None: A_ = t - 1 A_ = self.alphas_cumprod[t] A_ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one A_ = 1 - alpha_prod_t A_ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: A_ = self.betas[t] A_ = self.alphas[t] else: A_ = 1 - alpha_prod_t / alpha_prod_t_prev A_ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": A_ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": A_ = model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`''' " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: A_ = torch.clamp( UpperCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A_ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t A_ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A_ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise A_ = 0 if t > 0: A_ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=UpperCAmelCase , device=model_output.device ) A_ = self._get_variance( UpperCAmelCase , predicted_variance=UpperCAmelCase , prev_timestep=UpperCAmelCase , ) if self.variance_type == "fixed_small_log": A_ = variance elif self.variance_type == "learned_range": A_ = (0.5 * variance).exp() else: raise ValueError( f'''variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`''' " for the UnCLIPScheduler." ) A_ = variance * variance_noise A_ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=UpperCAmelCase , pred_original_sample=UpperCAmelCase ) def __A ( self : Optional[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : torch.IntTensor , ): # Make sure alphas_cumprod and timestep have same device and dtype as original_samples A_ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) A_ = timesteps.to(original_samples.device ) A_ = alphas_cumprod[timesteps] ** 0.5 A_ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): A_ = sqrt_alpha_prod.unsqueeze(-1 ) A_ = (1 - alphas_cumprod[timesteps]) ** 0.5 A_ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): A_ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) A_ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
329
1
'''simple docstring''' from collections import namedtuple import requests from lxml import html # type: ignore lowercase : Optional[int] = namedtuple("covid_data", "cases deaths recovered") def SCREAMING_SNAKE_CASE__ ( __A = "https://www.worldometers.info/coronavirus/" ) -> Dict: _snake_case = """//div[@class = \"maincounter-number\"]/span/text()""" return covid_data(*html.fromstring(requests.get(UpperCamelCase__ ).content ).xpath(UpperCamelCase__ ) ) lowercase : int = "Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}" print(fmt.format(*covid_stats()))
42
'''simple docstring''' from collections import namedtuple import requests from lxml import html # type: ignore __A =namedtuple('covid_data', 'cases deaths recovered') def _UpperCamelCase ( UpperCamelCase__ = "https://www.worldometers.info/coronavirus/" ): UpperCAmelCase__ : Union[str, Any] = """//div[@class = \"maincounter-number\"]/span/text()""" return covid_data(*html.fromstring(requests.get(UpperCamelCase__ ).content ).xpath(UpperCamelCase__ ) ) __A ='Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}' print(fmt.format(*covid_stats()))
163
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case_ = { '''configuration_roformer''': ['''ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoFormerConfig''', '''RoFormerOnnxConfig'''], '''tokenization_roformer''': ['''RoFormerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = ['''RoFormerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoFormerForCausalLM''', '''RoFormerForMaskedLM''', '''RoFormerForMultipleChoice''', '''RoFormerForQuestionAnswering''', '''RoFormerForSequenceClassification''', '''RoFormerForTokenClassification''', '''RoFormerLayer''', '''RoFormerModel''', '''RoFormerPreTrainedModel''', '''load_tf_weights_in_roformer''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRoFormerForCausalLM''', '''TFRoFormerForMaskedLM''', '''TFRoFormerForMultipleChoice''', '''TFRoFormerForQuestionAnswering''', '''TFRoFormerForSequenceClassification''', '''TFRoFormerForTokenClassification''', '''TFRoFormerLayer''', '''TFRoFormerModel''', '''TFRoFormerPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FlaxRoFormerForMaskedLM''', '''FlaxRoFormerForMultipleChoice''', '''FlaxRoFormerForQuestionAnswering''', '''FlaxRoFormerForSequenceClassification''', '''FlaxRoFormerForTokenClassification''', '''FlaxRoFormerModel''', '''FlaxRoFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
362
import numpy as np def snake_case__ ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[Any] ): '''simple docstring''' lowercase__ : List[Any] = int(np.ceil((x_end - xa) / h ) ) lowercase__ : Dict = np.zeros((n + 1,) ) lowercase__ : Dict = ya lowercase__ : Union[str, Any] = xa for k in range(SCREAMING_SNAKE_CASE_ ): lowercase__ : Optional[int] = f(SCREAMING_SNAKE_CASE_ , y[k] ) lowercase__ : Dict = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) lowercase__ : Any = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) lowercase__ : List[Any] = f(x + h , y[k] + h * ka ) lowercase__ : Tuple = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
216
0
# Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=0 ): '''simple docstring''' if name is None: _lowerCAmelCase : Union[str, Any] = None else: _lowerCAmelCase : Any = "." * max(0 , spaces - 2 ) + "# {:" + str(50 - spaces ) + "s}" _lowerCAmelCase : Tuple = fmt.format(_lowerCamelCase ) # Print and recurse (if needed). if isinstance(_lowerCamelCase , _lowerCamelCase ): if msg is not None: print(_lowerCamelCase ) for k in val.keys(): recursive_print(_lowerCamelCase , val[k] , spaces + 2 ) elif isinstance(_lowerCamelCase , torch.Tensor ): print(_lowerCamelCase , ":" , val.size() ) else: print(_lowerCamelCase , ":" , _lowerCamelCase ) def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Dict = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] _lowerCAmelCase : Optional[int] = (num_heads, hidden_size, num_splits) + input_shape[1:] _lowerCAmelCase : Dict = param.view(*_lowerCamelCase ) _lowerCAmelCase : Optional[Any] = param.transpose(0 , 2 ) _lowerCAmelCase : str = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] _lowerCAmelCase : Any = (num_heads, num_splits, hidden_size) + input_shape[1:] _lowerCAmelCase : List[Any] = param.view(*_lowerCamelCase ) _lowerCAmelCase : Tuple = param.transpose(0 , 1 ).contiguous() _lowerCAmelCase : Optional[int] = param.view(*_lowerCamelCase ) return param def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : str = {} # old versions did not store training args _lowerCAmelCase : Optional[Any] = input_state_dict.get("args" , _lowerCamelCase ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) _lowerCAmelCase : List[Any] = ds_args.padded_vocab_size _lowerCAmelCase : Dict = ds_args.max_position_embeddings _lowerCAmelCase : Any = ds_args.hidden_size _lowerCAmelCase : List[Any] = ds_args.num_layers _lowerCAmelCase : Any = ds_args.num_attention_heads _lowerCAmelCase : str = ds_args.ffn_hidden_size # pprint(config) # The number of heads. _lowerCAmelCase : int = config.n_head # The hidden_size per head. _lowerCAmelCase : List[Any] = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): _lowerCAmelCase : Tuple = input_state_dict["checkpoint_version"] else: _lowerCAmelCase : List[Any] = 0.0 # The model. _lowerCAmelCase : Tuple = input_state_dict["model"] # The language model. _lowerCAmelCase : List[Any] = model["language_model"] # The embeddings. _lowerCAmelCase : Any = lm["embedding"] # The word embeddings. _lowerCAmelCase : Union[str, Any] = embeddings["word_embeddings"]["weight"] # Truncate the embedding table to vocab_size rows. _lowerCAmelCase : Optional[int] = word_embeddings[: config.vocab_size, :] _lowerCAmelCase : int = word_embeddings # The position embeddings. _lowerCAmelCase : Any = embeddings["position_embeddings"]["weight"] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] _lowerCAmelCase : List[str] = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( F"pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match" ) # Store the position embeddings. _lowerCAmelCase : int = pos_embeddings # The transformer. _lowerCAmelCase : int = lm["transformer"] if "transformer" in lm.keys() else lm["encoder"] # The regex to extract layer names. _lowerCAmelCase : Optional[int] = re.compile(r"layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)" ) # The simple map of names for "automated" rules. _lowerCAmelCase : int = { "attention.dense": ".attn.c_proj.", "self_attention.dense": ".attn.c_proj.", "mlp.dense_h_to_4h": ".mlp.c_fc.", "mlp.dense_4h_to_h": ".mlp.c_proj.", } # Extract the layers. for key, val in transformer.items(): # Match the name. _lowerCAmelCase : Optional[Any] = layer_re.match(_lowerCamelCase ) # Stop if that's not a layer if m is None: break # The index of the layer. _lowerCAmelCase : Optional[Any] = int(m.group(1 ) ) # The name of the operation. _lowerCAmelCase : List[str] = m.group(2 ) # Is it a weight or a bias? _lowerCAmelCase : List[Any] = m.group(3 ) # The name of the layer. _lowerCAmelCase : str = F"transformer.h.{layer_idx}" # For layernorm(s), simply store the layer norm. if op_name.endswith("layernorm" ): _lowerCAmelCase : Optional[Any] = "ln_1" if op_name.startswith("input" ) else "ln_2" _lowerCAmelCase : List[Any] = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. _lowerCAmelCase : Any = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , _lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : Dict = causal_mask # Insert a "dummy" tensor for masked_bias. _lowerCAmelCase : Tuple = torch.tensor(-1e4 , dtype=torch.floataa ) _lowerCAmelCase : Tuple = masked_bias _lowerCAmelCase : List[Any] = fix_query_key_value_ordering(_lowerCamelCase , _lowerCamelCase , 3 , _lowerCamelCase , _lowerCamelCase ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. _lowerCAmelCase : Dict = out_val.transpose(0 , 1 ).contiguous() # Store. _lowerCAmelCase : List[Any] = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": _lowerCAmelCase : Optional[int] = fix_query_key_value_ordering(_lowerCamelCase , _lowerCamelCase , 3 , _lowerCamelCase , _lowerCamelCase ) # Store. No change of shape. _lowerCAmelCase : Tuple = out_val # Transpose the weights. elif weight_or_bias == "weight": _lowerCAmelCase : Any = megatron_to_transformers[op_name] _lowerCAmelCase : str = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": _lowerCAmelCase : Tuple = megatron_to_transformers[op_name] _lowerCAmelCase : Dict = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. _lowerCAmelCase : Dict = transformer["final_layernorm.weight"] _lowerCAmelCase : List[str] = transformer["final_layernorm.bias"] # For LM head, transformers' wants the matrix to weight embeddings. _lowerCAmelCase : Union[str, Any] = word_embeddings # It should be done! return output_state_dict def A ( ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("--print-checkpoint-structure" , action="store_true" ) parser.add_argument( "path_to_checkpoint" , type=_lowerCamelCase , help="Path to the checkpoint file (.zip archive or direct .pt file)" , ) parser.add_argument( "--config_file" , default="" , type=_lowerCamelCase , help="An optional config json file describing the pre-trained model." , ) _lowerCAmelCase : Union[str, Any] = parser.parse_args() # Extract the basename. _lowerCAmelCase : List[Any] = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(F"Extracting PyTorch state dictionary from {args.path_to_checkpoint}" ) if args.path_to_checkpoint.endswith(".zip" ): with zipfile.ZipFile(args.path_to_checkpoint , "r" ) as checkpoint: with checkpoint.open("release/mp_rank_00/model_optim_rng.pt" ) as pytorch_dict: _lowerCAmelCase : int = torch.load(_lowerCamelCase , map_location="cpu" ) else: _lowerCAmelCase : List[Any] = torch.load(args.path_to_checkpoint , map_location="cpu" ) _lowerCAmelCase : Dict = input_state_dict.get("args" , _lowerCamelCase ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: _lowerCAmelCase : Union[str, Any] = "gelu_fast" elif ds_args.openai_gelu: _lowerCAmelCase : int = "gelu_new" else: _lowerCAmelCase : Dict = "gelu" else: # in the very early days this used to be "gelu_new" _lowerCAmelCase : Dict = "gelu_new" # Spell out all parameters in case the defaults change. _lowerCAmelCase : Union[str, Any] = GPTaConfig( vocab_size=50_257 , n_positions=1_024 , n_embd=1_024 , n_layer=24 , n_head=16 , n_inner=4_096 , activation_function=_lowerCamelCase , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1e-5 , initializer_range=0.02 , summary_type="cls_index" , summary_use_proj=_lowerCamelCase , summary_activation=_lowerCamelCase , summary_proj_to_labels=_lowerCamelCase , summary_first_dropout=0.1 , scale_attn_weights=_lowerCamelCase , use_cache=_lowerCamelCase , bos_token_id=50_256 , eos_token_id=50_256 , ) else: _lowerCAmelCase : str = GPTaConfig.from_json_file(args.config_file ) _lowerCAmelCase : Dict = ["GPT2LMHeadModel"] # Convert. print("Converting" ) _lowerCAmelCase : Tuple = convert_megatron_checkpoint(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(_lowerCamelCase , _lowerCamelCase ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: _lowerCAmelCase : Optional[int] = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": _lowerCAmelCase : Optional[int] = "gpt2" elif tokenizer_type == "PretrainedFromHF": _lowerCAmelCase : Union[str, Any] = ds_args.tokenizer_name_or_path else: raise ValueError(F"Unrecognized tokenizer_type {tokenizer_type}" ) else: _lowerCAmelCase : str = "gpt2" _lowerCAmelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCamelCase ) _lowerCAmelCase : Dict = type(_lowerCamelCase ).__name__ _lowerCAmelCase : List[str] = tokenizer_class # Store the config to file. print("Saving config" ) config.save_pretrained(_lowerCamelCase ) # Save tokenizer based on args print(F"Adding {tokenizer_class} tokenizer files" ) tokenizer.save_pretrained(_lowerCamelCase ) # Store the state_dict to file. _lowerCAmelCase : List[str] = os.path.join(_lowerCamelCase , "pytorch_model.bin" ) print(F"Saving checkpoint to \"{output_checkpoint_file}\"" ) torch.save(_lowerCamelCase , _lowerCamelCase ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
36
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() _A = logging.get_logger(__name__) def __UpperCamelCase ( _A ): lowerCAmelCase_ = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: lowerCAmelCase_ = 192 lowerCAmelCase_ = 768 lowerCAmelCase_ = 12 lowerCAmelCase_ = 3 lowerCAmelCase_ = [800, 1333] lowerCAmelCase_ = False elif yolos_name == "yolos_s_dWr": lowerCAmelCase_ = 330 lowerCAmelCase_ = 14 lowerCAmelCase_ = 6 lowerCAmelCase_ = 1320 elif "yolos_s" in yolos_name: lowerCAmelCase_ = 384 lowerCAmelCase_ = 1536 lowerCAmelCase_ = 12 lowerCAmelCase_ = 6 elif "yolos_b" in yolos_name: lowerCAmelCase_ = [800, 1344] lowerCAmelCase_ = 91 lowerCAmelCase_ = '''huggingface/label-files''' lowerCAmelCase_ = '''coco-detection-id2label.json''' lowerCAmelCase_ = json.load(open(hf_hub_download(_A , _A , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase_ = {int(_A ): v for k, v in idalabel.items()} lowerCAmelCase_ = idalabel lowerCAmelCase_ = {v: k for k, v in idalabel.items()} return config def __UpperCamelCase ( _A , _A , _A = False ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase_ = state_dict.pop(f"blocks.{i}.attn.qkv.weight" ) lowerCAmelCase_ = state_dict.pop(f"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase_ = in_proj_weight[: config.hidden_size, :] lowerCAmelCase_ = in_proj_bias[: config.hidden_size] lowerCAmelCase_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase_ = in_proj_weight[-config.hidden_size :, :] lowerCAmelCase_ = in_proj_bias[-config.hidden_size :] def __UpperCamelCase ( _A ): if "backbone" in name: lowerCAmelCase_ = name.replace('''backbone''' , '''vit''' ) if "cls_token" in name: lowerCAmelCase_ = name.replace('''cls_token''' , '''embeddings.cls_token''' ) if "det_token" in name: lowerCAmelCase_ = name.replace('''det_token''' , '''embeddings.detection_tokens''' ) if "mid_pos_embed" in name: lowerCAmelCase_ = name.replace('''mid_pos_embed''' , '''encoder.mid_position_embeddings''' ) if "pos_embed" in name: lowerCAmelCase_ = name.replace('''pos_embed''' , '''embeddings.position_embeddings''' ) if "patch_embed.proj" in name: lowerCAmelCase_ = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "blocks" in name: lowerCAmelCase_ = name.replace('''blocks''' , '''encoder.layer''' ) if "attn.proj" in name: lowerCAmelCase_ = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowerCAmelCase_ = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowerCAmelCase_ = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowerCAmelCase_ = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowerCAmelCase_ = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowerCAmelCase_ = name.replace('''mlp.fc2''' , '''output.dense''' ) if "class_embed" in name: lowerCAmelCase_ = name.replace('''class_embed''' , '''class_labels_classifier''' ) if "bbox_embed" in name: lowerCAmelCase_ = name.replace('''bbox_embed''' , '''bbox_predictor''' ) if "vit.norm" in name: lowerCAmelCase_ = name.replace('''vit.norm''' , '''vit.layernorm''' ) return name def __UpperCamelCase ( _A , _A ): for key in orig_state_dict.copy().keys(): lowerCAmelCase_ = orig_state_dict.pop(_A ) if "qkv" in key: lowerCAmelCase_ = key.split('''.''' ) lowerCAmelCase_ = int(key_split[2] ) lowerCAmelCase_ = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: lowerCAmelCase_ = val[:dim, :] lowerCAmelCase_ = val[ dim : dim * 2, : ] lowerCAmelCase_ = val[-dim:, :] else: lowerCAmelCase_ = val[:dim] lowerCAmelCase_ = val[dim : dim * 2] lowerCAmelCase_ = val[-dim:] else: lowerCAmelCase_ = val return orig_state_dict def __UpperCamelCase ( ): lowerCAmelCase_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase_ = Image.open(requests.get(_A , stream=_A ).raw ) return im @torch.no_grad() def __UpperCamelCase ( _A , _A , _A , _A = False ): lowerCAmelCase_ = get_yolos_config(_A ) # load original state_dict lowerCAmelCase_ = torch.load(_A , map_location='''cpu''' )['''model'''] # load 🤗 model lowerCAmelCase_ = YolosForObjectDetection(_A ) model.eval() lowerCAmelCase_ = convert_state_dict(_A , _A ) model.load_state_dict(_A ) # Check outputs on an image, prepared by YolosImageProcessor lowerCAmelCase_ = 800 if yolos_name != '''yolos_ti''' else 512 lowerCAmelCase_ = YolosImageProcessor(format='''coco_detection''' , size=_A ) lowerCAmelCase_ = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowerCAmelCase_ = model(**_A ) lowerCAmelCase_ , lowerCAmelCase_ = outputs.logits, outputs.pred_boxes lowerCAmelCase_ , lowerCAmelCase_ = None, None if yolos_name == "yolos_ti": lowerCAmelCase_ = torch.tensor( [[-3_9.5_0_2_2, -1_1.9_8_2_0, -1_7.6_8_8_8], [-2_9.9_5_7_4, -9.9_7_6_9, -1_7.7_6_9_1], [-4_2.3_2_8_1, -2_0.7_2_0_0, -3_0.6_2_9_4]] ) lowerCAmelCase_ = torch.tensor( [[0.4_0_2_1, 0.0_8_3_6, 0.7_9_7_9], [0.0_1_8_4, 0.2_6_0_9, 0.0_3_6_4], [0.1_7_8_1, 0.2_0_0_4, 0.2_0_9_5]] ) elif yolos_name == "yolos_s_200_pre": lowerCAmelCase_ = torch.tensor( [[-2_4.0_2_4_8, -1_0.3_0_2_4, -1_4.8_2_9_0], [-4_2.0_3_9_2, -1_6.8_2_0_0, -2_7.4_3_3_4], [-2_7.2_7_4_3, -1_1.8_1_5_4, -1_8.7_1_4_8]] ) lowerCAmelCase_ = torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] ) elif yolos_name == "yolos_s_300_pre": lowerCAmelCase_ = torch.tensor( [[-3_6.2_2_2_0, -1_4.4_3_8_5, -2_3.5_4_5_7], [-3_5.6_9_7_0, -1_4.7_5_8_3, -2_1.3_9_3_5], [-3_1.5_9_3_9, -1_3.6_0_4_2, -1_6.8_0_4_9]] ) lowerCAmelCase_ = torch.tensor( [[0.7_6_1_4, 0.2_3_1_6, 0.4_7_2_8], [0.7_1_6_8, 0.4_4_9_5, 0.3_8_5_5], [0.4_9_9_6, 0.1_4_6_6, 0.9_9_9_6]] ) elif yolos_name == "yolos_s_dWr": lowerCAmelCase_ = torch.tensor( [[-4_2.8_6_6_8, -2_4.1_0_4_9, -4_1.1_6_9_0], [-3_4.7_4_5_6, -1_4.1_2_7_4, -2_4.9_1_9_4], [-3_3.7_8_9_8, -1_2.1_9_4_6, -2_5.6_4_9_5]] ) lowerCAmelCase_ = torch.tensor( [[0.5_5_8_7, 0.2_7_7_3, 0.0_6_0_5], [0.5_0_0_4, 0.3_0_1_4, 0.9_9_9_4], [0.4_9_9_9, 0.1_5_4_8, 0.9_9_9_4]] ) elif yolos_name == "yolos_base": lowerCAmelCase_ = torch.tensor( [[-4_0.6_0_6_4, -2_4.3_0_8_4, -3_2.6_4_4_7], [-5_5.1_9_9_0, -3_0.7_7_1_9, -3_5.5_8_7_7], [-5_1.4_3_1_1, -3_3.3_5_0_7, -3_5.6_4_6_2]] ) lowerCAmelCase_ = torch.tensor( [[0.5_5_5_5, 0.2_7_9_4, 0.0_6_5_5], [0.9_0_4_9, 0.2_6_6_4, 0.1_8_9_4], [0.9_1_8_3, 0.1_9_8_4, 0.1_6_3_5]] ) else: raise ValueError(f"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , _A , atol=1E-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , _A , atol=1E-4 ) Path(_A ).mkdir(exist_ok=_A ) print(f"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_A ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(_A ) if push_to_hub: lowerCAmelCase_ = { '''yolos_ti''': '''yolos-tiny''', '''yolos_s_200_pre''': '''yolos-small''', '''yolos_s_300_pre''': '''yolos-small-300''', '''yolos_s_dWr''': '''yolos-small-dwr''', '''yolos_base''': '''yolos-base''', } print('''Pushing to the hub...''' ) lowerCAmelCase_ = model_mapping[yolos_name] image_processor.push_to_hub(_A , organization='''hustvl''' ) model.push_to_hub(_A , organization='''hustvl''' ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _A = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
278
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor snake_case_ : Union[str, Any] = logging.get_logger(__name__) class lowercase__ ( lowercase ): def __init__( self : Optional[int] ,*lowerCamelCase__ : Dict ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' warnings.warn( 'The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ImageGPTImageProcessor instead.' ,lowerCamelCase__ ,) super().__init__(*lowerCamelCase__ ,**lowerCamelCase__ )
371
'''simple docstring''' import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # Initialise PyTorch model _UpperCamelCase : Any = LxmertConfig.from_json_file(UpperCAmelCase_ ) print(f'Building PyTorch model from configuration: {config}' ) _UpperCamelCase : int = LxmertForPreTraining(UpperCAmelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , UpperCAmelCase_ ) if __name__ == "__main__": snake_case_ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) snake_case_ : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
236
0
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class __lowercase ( _UpperCamelCase ): '''simple docstring''' def __init__( self , *_UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , **_UpperCAmelCase ): super().__init__(*_UpperCAmelCase , **_UpperCAmelCase ) __a : int = eval_examples __a : int = post_process_function def _lowerCamelCase ( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase = "eval" ): __a : str = self.eval_dataset if eval_dataset is None else eval_dataset __a : Union[str, Any] = self.get_eval_dataloader(_UpperCAmelCase ) __a : int = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. __a : Any = self.compute_metrics __a : Union[str, Any] = None __a : Optional[int] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __a : Optional[Any] = time.time() try: __a : Union[str, Any] = eval_loop( _UpperCAmelCase , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_UpperCAmelCase , metric_key_prefix=_UpperCAmelCase , ) finally: __a : str = compute_metrics __a : List[str] = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _UpperCAmelCase , _UpperCAmelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default __a : Union[str, Any] = self.post_process_function(_UpperCAmelCase , _UpperCAmelCase , output.predictions ) __a : Dict = self.compute_metrics(_UpperCAmelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __a : List[str] = metrics.pop(_UpperCAmelCase ) metrics.update(output.metrics ) else: __a : Union[str, Any] = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_UpperCAmelCase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) __a : Union[str, Any] = self.callback_handler.on_evaluate(self.args , self.state , self.control , _UpperCAmelCase ) return metrics def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase = "test" ): __a : Tuple = self.get_test_dataloader(_UpperCAmelCase ) # Temporarily disable metric computation, we will do it in the loop here. __a : List[str] = self.compute_metrics __a : Optional[Any] = None __a : Optional[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __a : Tuple = time.time() try: __a : Dict = eval_loop( _UpperCAmelCase , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_UpperCAmelCase , metric_key_prefix=_UpperCAmelCase , ) finally: __a : Optional[Any] = compute_metrics __a : Any = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _UpperCAmelCase , _UpperCAmelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output __a : Dict = self.post_process_function(_UpperCAmelCase , _UpperCAmelCase , output.predictions , '''predict''' ) __a : List[Any] = self.compute_metrics(_UpperCAmelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __a : int = metrics.pop(_UpperCAmelCase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_UpperCAmelCase )
160
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A = { '''configuration_ctrl''': ['''CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CTRLConfig'''], '''tokenization_ctrl''': ['''CTRLTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''CTRL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CTRLForSequenceClassification''', '''CTRLLMHeadModel''', '''CTRLModel''', '''CTRLPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFCTRLForSequenceClassification''', '''TFCTRLLMHeadModel''', '''TFCTRLModel''', '''TFCTRLPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
160
1
"""simple docstring""" _lowerCAmelCase : Dict = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
340
"""simple docstring""" def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> float: '''simple docstring''' _validate_point(_lowerCamelCase ) _validate_point(_lowerCamelCase ) if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError("Both points must be in the same n-dimensional space" ) return float(sum(abs(a - b ) for a, b in zip(_lowerCamelCase , _lowerCamelCase ) ) ) def lowerCamelCase_( _lowerCamelCase ) -> None: '''simple docstring''' if point: if isinstance(_lowerCamelCase , _lowerCamelCase ): for item in point: if not isinstance(_lowerCamelCase , (int, float) ): _lowerCamelCase : Dict = ( "Expected a list of numbers as input, found " F"""{type(_lowerCamelCase ).__name__}""" ) raise TypeError(_lowerCamelCase ) else: _lowerCamelCase : Optional[int] = F"""Expected a list of numbers as input, found {type(_lowerCamelCase ).__name__}""" raise TypeError(_lowerCamelCase ) else: raise ValueError("Missing an input" ) def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> float: '''simple docstring''' _validate_point(_lowerCamelCase ) _validate_point(_lowerCamelCase ) if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError("Both points must be in the same n-dimensional space" ) return float(sum(abs(x - y ) for x, y in zip(_lowerCamelCase , _lowerCamelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
340
1
import random from typing import Any def a_ ( __lowercase : list ) -> list[Any]: for _ in range(len(__lowercase ) ): _snake_case = random.randint(0 , len(__lowercase ) - 1 ) _snake_case = random.randint(0 , len(__lowercase ) - 1 ) _snake_case , _snake_case = data[b], data[a] return data if __name__ == "__main__": _lowerCamelCase : Tuple = [0, 1, 2, 3, 4, 5, 6, 7] _lowerCamelCase : Union[str, Any] = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
282
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class SCREAMING_SNAKE_CASE__ : '''simple docstring''' @property def A ( self : List[str] ): '''simple docstring''' return self.get_dummy_input() @property def A ( self : Any ): '''simple docstring''' if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def A ( self : Union[str, Any] , lowercase : Any=True , lowercase : List[Any]=False , lowercase : List[str]=False , lowercase : Dict=False , ): '''simple docstring''' _snake_case = 4 _snake_case = 32 _snake_case = (32, 32) _snake_case = torch.manual_seed(0 ) _snake_case = torch.device(lowercase ) _snake_case = (batch_size, num_channels) + sizes _snake_case = randn_tensor(lowercase , generator=lowercase , device=lowercase ) _snake_case = {'hidden_states': hidden_states} if include_temb: _snake_case = 128 _snake_case = randn_tensor((batch_size, temb_channels) , generator=lowercase , device=lowercase ) if include_res_hidden_states_tuple: _snake_case = torch.manual_seed(1 ) _snake_case = (randn_tensor(lowercase , generator=lowercase , device=lowercase ),) if include_encoder_hidden_states: _snake_case = floats_tensor((batch_size, 32, 32) ).to(lowercase ) if include_skip_sample: _snake_case = randn_tensor(((batch_size, 3) + sizes) , generator=lowercase , device=lowercase ) return dummy_input def A ( self : Any ): '''simple docstring''' _snake_case = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": _snake_case = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) _snake_case = self.dummy_input return init_dict, inputs_dict def A ( self : Dict , lowercase : Optional[int] ): '''simple docstring''' _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowercase ) unet_block.to(lowercase ) unet_block.eval() with torch.no_grad(): _snake_case = unet_block(**lowercase ) if isinstance(lowercase , lowercase ): _snake_case = output[0] self.assertEqual(output.shape , self.output_shape ) _snake_case = output[0, -1, -3:, -3:] _snake_case = torch.tensor(lowercase ).to(lowercase ) assert torch_all_close(output_slice.flatten() , lowercase , atol=5E-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def A ( self : Dict ): '''simple docstring''' _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowercase ) model.to(lowercase ) model.train() _snake_case = model(**lowercase ) if isinstance(lowercase , lowercase ): _snake_case = output[0] _snake_case = torch.device(lowercase ) _snake_case = randn_tensor(output.shape , device=lowercase ) _snake_case = torch.nn.functional.mse_loss(lowercase , lowercase ) loss.backward()
282
1
from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ = 9, 14 # noqa: F841 UpperCamelCase__ = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] UpperCamelCase__ = defaultdict(lowerCamelCase__ ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) UpperCamelCase__ = mst(lowerCamelCase__ ) UpperCamelCase__ = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: UpperCamelCase__ = tuple(answer[:2] ) UpperCamelCase__ = tuple(edge[::-1] ) assert edge in result or reverse in result
350
from __future__ import annotations def __magic_name__ ( __a : list[list[int]] ): '''simple docstring''' for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(__a ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(__a ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
178
0
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import FEATURE_EXTRACTOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor, ChineseCLIPProcessor @require_vision class lowercase ( unittest.TestCase): def a_ ( self : Tuple ): """simple docstring""" A_ : Tuple = tempfile.mkdtemp() A_ : Union[str, Any] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''的''', '''价''', '''格''', '''是''', '''15''', '''便''', '''alex''', '''##andra''', ''',''', '''。''', '''-''', '''t''', '''shirt''', ] A_ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) A_ : int = { '''do_resize''': True, '''size''': {'''height''': 2_24, '''width''': 2_24}, '''do_center_crop''': True, '''crop_size''': {'''height''': 18, '''width''': 18}, '''do_normalize''': True, '''image_mean''': [0.48145466, 0.4578275, 0.40821073], '''image_std''': [0.26862954, 0.26130258, 0.27577711], '''do_convert_rgb''': True, } A_ : List[Any] = os.path.join(self.tmpdirname , _lowerCamelCase ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(_lowerCamelCase , _lowerCamelCase ) def a_ ( self : Tuple , **_lowerCamelCase : List[str] ): """simple docstring""" return BertTokenizer.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def a_ ( self : str , **_lowerCamelCase : List[Any] ): """simple docstring""" return BertTokenizerFast.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def a_ ( self : int , **_lowerCamelCase : Optional[Any] ): """simple docstring""" return ChineseCLIPImageProcessor.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def a_ ( self : Optional[Any] ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def a_ ( self : Union[str, Any] ): """simple docstring""" A_ : Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] A_ : Optional[int] = [Image.fromarray(np.moveaxis(_lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def a_ ( self : Any ): """simple docstring""" A_ : List[Any] = self.get_tokenizer() A_ : Optional[int] = self.get_rust_tokenizer() A_ : Optional[Any] = self.get_image_processor() A_ : Union[str, Any] = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) A_ : Optional[Any] = ChineseCLIPProcessor.from_pretrained(self.tmpdirname , use_fast=_lowerCamelCase ) A_ : Tuple = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) A_ : Optional[Any] = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , _lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , _lowerCamelCase ) def a_ ( self : int ): """simple docstring""" A_ : Union[str, Any] = ChineseCLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) A_ : Tuple = self.get_tokenizer(cls_token='''(CLS)''' , sep_token='''(SEP)''' ) A_ : List[str] = self.get_image_processor(do_normalize=_lowerCamelCase ) A_ : Union[str, Any] = ChineseCLIPProcessor.from_pretrained( self.tmpdirname , cls_token='''(CLS)''' , sep_token='''(SEP)''' , do_normalize=_lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCamelCase ) def a_ ( self : Tuple ): """simple docstring""" A_ : Optional[int] = self.get_image_processor() A_ : int = self.get_tokenizer() A_ : Optional[Any] = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : Optional[Any] = self.prepare_image_inputs() A_ : Dict = image_processor(_lowerCamelCase , return_tensors='''np''' ) A_ : Optional[int] = processor(images=_lowerCamelCase , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a_ ( self : Tuple ): """simple docstring""" A_ : Tuple = self.get_image_processor() A_ : Any = self.get_tokenizer() A_ : Dict = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : List[Any] = '''Alexandra,T-shirt的价格是15便士。''' A_ : Any = processor(text=_lowerCamelCase ) A_ : Any = tokenizer(_lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def a_ ( self : Optional[Any] ): """simple docstring""" A_ : Any = self.get_image_processor() A_ : Tuple = self.get_tokenizer() A_ : Any = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : Any = '''Alexandra,T-shirt的价格是15便士。''' A_ : Union[str, Any] = self.prepare_image_inputs() A_ : Optional[int] = processor(text=_lowerCamelCase , images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(_lowerCamelCase ): processor() def a_ ( self : Union[str, Any] ): """simple docstring""" A_ : int = self.get_image_processor() A_ : int = self.get_tokenizer() A_ : Dict = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : str = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A_ : Optional[int] = processor.batch_decode(_lowerCamelCase ) A_ : List[Any] = tokenizer.batch_decode(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def a_ ( self : Any ): """simple docstring""" A_ : str = self.get_image_processor() A_ : Any = self.get_tokenizer() A_ : Any = ChineseCLIPProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : Tuple = '''Alexandra,T-shirt的价格是15便士。''' A_ : Tuple = self.prepare_image_inputs() A_ : Dict = processor(text=_lowerCamelCase , images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
167
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class lowercase ( __UpperCAmelCase): __lowerCAmelCase : str = ["""image_processor""", """tokenizer"""] __lowerCAmelCase : Optional[Any] = """OwlViTImageProcessor""" __lowerCAmelCase : Any = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : Union[str, Any] , _lowerCamelCase : str=None , _lowerCamelCase : Tuple=None , **_lowerCamelCase : List[Any] ): """simple docstring""" A_ : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _lowerCamelCase , ) A_ : List[Any] = kwargs.pop('''feature_extractor''' ) A_ : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_lowerCamelCase , _lowerCamelCase ) def __call__( self : Optional[int] , _lowerCamelCase : List[Any]=None , _lowerCamelCase : Dict=None , _lowerCamelCase : int=None , _lowerCamelCase : str="max_length" , _lowerCamelCase : List[Any]="np" , **_lowerCamelCase : Optional[int] ): """simple docstring""" if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''' ) if text is not None: if isinstance(_lowerCamelCase , _lowerCamelCase ) or (isinstance(_lowerCamelCase , _lowerCamelCase ) and not isinstance(text[0] , _lowerCamelCase )): A_ : List[str] = [self.tokenizer(_lowerCamelCase , padding=_lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase )] elif isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(text[0] , _lowerCamelCase ): A_ : Optional[int] = [] # Maximum number of queries across batch A_ : Any = max([len(_lowerCamelCase ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(_lowerCamelCase ) != max_num_queries: A_ : Optional[int] = t + [''' '''] * (max_num_queries - len(_lowerCamelCase )) A_ : Tuple = self.tokenizer(_lowerCamelCase , padding=_lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase ) encodings.append(_lowerCamelCase ) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''' ) if return_tensors == "np": A_ : Union[str, Any] = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) A_ : Dict = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp A_ : List[Any] = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) A_ : Any = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch A_ : Any = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0 ) A_ : Union[str, Any] = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf A_ : Tuple = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) A_ : str = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) else: raise ValueError('''Target return tensor type could not be returned''' ) A_ : Any = BatchEncoding() A_ : Optional[Any] = input_ids A_ : str = attention_mask if query_images is not None: A_ : Union[str, Any] = BatchEncoding() A_ : Optional[Any] = self.image_processor( _lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase ).pixel_values A_ : Dict = query_pixel_values if images is not None: A_ : int = self.image_processor(_lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase ) if text is not None and images is not None: A_ : str = image_features.pixel_values return encoding elif query_images is not None and images is not None: A_ : Union[str, Any] = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCamelCase ) , tensor_type=_lowerCamelCase ) def a_ ( self : Optional[Any] , *_lowerCamelCase : int , **_lowerCamelCase : Dict ): """simple docstring""" return self.image_processor.post_process(*_lowerCamelCase , **_lowerCamelCase ) def a_ ( self : Optional[Any] , *_lowerCamelCase : Optional[Any] , **_lowerCamelCase : Dict ): """simple docstring""" return self.image_processor.post_process_object_detection(*_lowerCamelCase , **_lowerCamelCase ) def a_ ( self : List[Any] , *_lowerCamelCase : List[str] , **_lowerCamelCase : Optional[int] ): """simple docstring""" return self.image_processor.post_process_image_guided_detection(*_lowerCamelCase , **_lowerCamelCase ) def a_ ( self : str , *_lowerCamelCase : Tuple , **_lowerCamelCase : List[Any] ): """simple docstring""" return self.tokenizer.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def a_ ( self : Dict , *_lowerCamelCase : Any , **_lowerCamelCase : Union[str, Any] ): """simple docstring""" return self.tokenizer.decode(*_lowerCamelCase , **_lowerCamelCase ) @property def a_ ( self : List[str] ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _lowerCamelCase , ) return self.image_processor_class @property def a_ ( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _lowerCamelCase , ) return self.image_processor
167
1
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def __lowercase ( snake_case_ : str = "isbn/0140328726" ) ->dict: '''simple docstring''' __A : int = olid.strip().strip('''/''' ) # Remove leading/trailing whitespace & slashes if new_olid.count('''/''' ) != 1: __A : Tuple = F"""{olid} is not a valid Open Library olid""" raise ValueError(snake_case_ ) return requests.get(F"""https://openlibrary.org/{new_olid}.json""" ).json() def __lowercase ( snake_case_ : dict ) ->dict: '''simple docstring''' __A : str = { '''title''': '''Title''', '''publish_date''': '''Publish date''', '''authors''': '''Authors''', '''number_of_pages''': '''Number of pages:''', '''first_sentence''': '''First sentence''', '''isbn_10''': '''ISBN (10)''', '''isbn_13''': '''ISBN (13)''', } __A : str = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} __A : Dict = [ get_openlibrary_data(author['''key'''] )['''name'''] for author in data['''Authors'''] ] __A : str = data['''First sentence''']['''value'''] for key, value in data.items(): if isinstance(snake_case_ ,snake_case_ ): __A : Optional[Any] = ''', '''.join(snake_case_ ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: a_ = input("""\nEnter the ISBN code to search (or 'quit' to stop): """).strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(f'''Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.''') continue print(f'''\nSearching Open Library for ISBN: {isbn}...\n''') try: a_ = summarize_book(get_openlibrary_data(f'''isbn/{isbn}''')) print("""\n""".join(f'''{key}: {value}''' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(f'''Sorry, there are no results for ISBN: {isbn}.''')
364
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ = { """configuration_bloom""": ["""BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BloomConfig""", """BloomOnnxConfig"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ["""BloomTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST""", """BloomForCausalLM""", """BloomModel""", """BloomPreTrainedModel""", """BloomForSequenceClassification""", """BloomForTokenClassification""", """BloomForQuestionAnswering""", ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
291
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { '''google/canine-s''': '''https://huggingface.co/google/canine-s/resolve/main/config.json''', # See all CANINE models at https://huggingface.co/models?filter=canine } class _snake_case ( a__ ): snake_case__ = "canine" def __init__( self : Union[str, Any] , UpperCAmelCase : Optional[int]=768 , UpperCAmelCase : Tuple=12 , UpperCAmelCase : Optional[int]=12 , UpperCAmelCase : Union[str, Any]=3072 , UpperCAmelCase : Optional[Any]="gelu" , UpperCAmelCase : Union[str, Any]=0.1 , UpperCAmelCase : Tuple=0.1 , UpperCAmelCase : List[str]=16384 , UpperCAmelCase : List[str]=16 , UpperCAmelCase : int=0.0_2 , UpperCAmelCase : List[Any]=1E-12 , UpperCAmelCase : Optional[Any]=0 , UpperCAmelCase : Optional[int]=0XE_000 , UpperCAmelCase : Dict=0XE_001 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Any=4 , UpperCAmelCase : List[Any]=8 , UpperCAmelCase : Optional[int]=16384 , UpperCAmelCase : Dict=128 , **UpperCAmelCase : List[str] , ): super().__init__(pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) __lowerCamelCase : Any = max_position_embeddings __lowerCamelCase : Dict = hidden_size __lowerCamelCase : Any = num_hidden_layers __lowerCamelCase : Tuple = num_attention_heads __lowerCamelCase : Any = intermediate_size __lowerCamelCase : Tuple = hidden_act __lowerCamelCase : Tuple = hidden_dropout_prob __lowerCamelCase : List[str] = attention_probs_dropout_prob __lowerCamelCase : str = initializer_range __lowerCamelCase : Dict = type_vocab_size __lowerCamelCase : List[str] = layer_norm_eps # Character config: __lowerCamelCase : List[str] = downsampling_rate __lowerCamelCase : List[Any] = upsampling_kernel_size __lowerCamelCase : List[str] = num_hash_functions __lowerCamelCase : Dict = num_hash_buckets __lowerCamelCase : str = local_transformer_stride
135
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A = {'''configuration_mra''': ['''MRA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MraConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ '''MRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MraForMaskedLM''', '''MraForMultipleChoice''', '''MraForQuestionAnswering''', '''MraForSequenceClassification''', '''MraForTokenClassification''', '''MraLayer''', '''MraModel''', '''MraPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
135
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self , __magic_name__ ) -> List[Any]: '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): snake_case_ : Union[str, Any] = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(__magic_name__ ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : Any = '''sshleifer/tiny-gpt2''' snake_case_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : Optional[int] = PyTorchBenchmark(__magic_name__ ) snake_case_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = '''sgugger/tiny-distilbert-classification''' snake_case_ : List[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , only_pretrain_model=__magic_name__ , ) snake_case_ : Optional[Any] = PyTorchBenchmark(__magic_name__ ) snake_case_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Any = '''sshleifer/tiny-gpt2''' snake_case_ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , torchscript=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : Dict = PyTorchBenchmark(__magic_name__ ) snake_case_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : Tuple = '''sshleifer/tiny-gpt2''' snake_case_ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , fpaa=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : List[str] = PyTorchBenchmark(__magic_name__ ) snake_case_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = '''sshleifer/tiny-gpt2''' snake_case_ : int = AutoConfig.from_pretrained(__magic_name__ ) # set architectures equal to `None` snake_case_ : List[str] = None snake_case_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : List[str] = PyTorchBenchmark(__magic_name__ , configs=[config] ) snake_case_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = '''sshleifer/tiny-gpt2''' snake_case_ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : int = PyTorchBenchmark(__magic_name__ ) snake_case_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Can\'t do half precision''' ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = '''sshleifer/tiny-gpt2''' snake_case_ : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , fpaa=__magic_name__ , multi_process=__magic_name__ , ) snake_case_ : Dict = PyTorchBenchmark(__magic_name__ ) snake_case_ : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Any = '''sshleifer/tiny-gpt2''' snake_case_ : Dict = AutoConfig.from_pretrained(__magic_name__ ) snake_case_ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : List[Any] = PyTorchBenchmark(__magic_name__ , configs=[config] ) snake_case_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Dict = '''sshleifer/tinier_bart''' snake_case_ : Tuple = AutoConfig.from_pretrained(__magic_name__ ) snake_case_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : Optional[int] = PyTorchBenchmark(__magic_name__ , configs=[config] ) snake_case_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = '''sshleifer/tiny-gpt2''' snake_case_ : str = AutoConfig.from_pretrained(__magic_name__ ) snake_case_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : str = PyTorchBenchmark(__magic_name__ , configs=[config] ) snake_case_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = '''sshleifer/tinier_bart''' snake_case_ : Optional[int] = AutoConfig.from_pretrained(__magic_name__ ) snake_case_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__magic_name__ , ) snake_case_ : Optional[Any] = PyTorchBenchmark(__magic_name__ , configs=[config] ) snake_case_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : Dict = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , save_to_csv=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(__magic_name__ , '''inf_time.csv''' ) , train_memory_csv_file=os.path.join(__magic_name__ , '''train_mem.csv''' ) , inference_memory_csv_file=os.path.join(__magic_name__ , '''inf_mem.csv''' ) , train_time_csv_file=os.path.join(__magic_name__ , '''train_time.csv''' ) , env_info_csv_file=os.path.join(__magic_name__ , '''env.csv''' ) , multi_process=__magic_name__ , ) snake_case_ : Union[str, Any] = PyTorchBenchmark(__magic_name__ ) benchmark.run() self.assertTrue(Path(os.path.join(__magic_name__ , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(__magic_name__ , '''train_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(__magic_name__ , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(__magic_name__ , '''train_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(__magic_name__ , '''env.csv''' ) ).exists() ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Any = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(__magic_name__ ): self.assertTrue(hasattr(__magic_name__ , '''sequential''' ) ) self.assertTrue(hasattr(__magic_name__ , '''cumulative''' ) ) self.assertTrue(hasattr(__magic_name__ , '''current''' ) ) self.assertTrue(hasattr(__magic_name__ , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__magic_name__ , inference=__magic_name__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(__magic_name__ , '''log.txt''' ) , log_print=__magic_name__ , trace_memory_line_by_line=__magic_name__ , multi_process=__magic_name__ , ) snake_case_ : int = PyTorchBenchmark(__magic_name__ ) snake_case_ : Union[str, Any] = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(__magic_name__ , '''log.txt''' ) ).exists() )
279
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar lowerCAmelCase_ = TypeVar('''KEY''') lowerCAmelCase_ = TypeVar('''VAL''') @dataclass(frozen=_a, slots=_a ) class __lowerCAmelCase ( Generic[KEY, VAL] ): lowerCamelCase_ : KEY lowerCamelCase_ : VAL class __lowerCAmelCase ( _Item ): def __init__(self ) -> None: '''simple docstring''' super().__init__(__magic_name__ , __magic_name__ ) def __bool__(self ) -> bool: '''simple docstring''' return False lowerCAmelCase_ = _DeletedItem() class __lowerCAmelCase ( MutableMapping[KEY, VAL] ): def __init__(self , __magic_name__ = 8 , __magic_name__ = 0.75 ) -> None: '''simple docstring''' snake_case_ : List[Any] = initial_block_size snake_case_ : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 snake_case_ : List[str] = capacity_factor snake_case_ : int = 0 def lowerCamelCase (self , __magic_name__ ) -> int: '''simple docstring''' return hash(__magic_name__ ) % len(self._buckets ) def lowerCamelCase (self , __magic_name__ ) -> int: '''simple docstring''' return (ind + 1) % len(self._buckets ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ ) -> bool: '''simple docstring''' snake_case_ : Optional[int] = self._buckets[ind] if not stored: snake_case_ : Optional[Any] = _Item(__magic_name__ , __magic_name__ ) self._len += 1 return True elif stored.key == key: snake_case_ : List[Any] = _Item(__magic_name__ , __magic_name__ ) return True else: return False def lowerCamelCase (self ) -> bool: '''simple docstring''' snake_case_ : Any = len(self._buckets ) * self._capacity_factor return len(self ) >= int(__magic_name__ ) def lowerCamelCase (self ) -> bool: '''simple docstring''' if len(self._buckets ) <= self._initial_block_size: return False snake_case_ : int = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def lowerCamelCase (self , __magic_name__ ) -> None: '''simple docstring''' snake_case_ : List[str] = self._buckets snake_case_ : int = [None] * new_size snake_case_ : Optional[int] = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def lowerCamelCase (self ) -> None: '''simple docstring''' self._resize(len(self._buckets ) * 2 ) def lowerCamelCase (self ) -> None: '''simple docstring''' self._resize(len(self._buckets ) // 2 ) def lowerCamelCase (self , __magic_name__ ) -> Iterator[int]: '''simple docstring''' snake_case_ : Dict = self._get_bucket_index(__magic_name__ ) for _ in range(len(self._buckets ) ): yield ind snake_case_ : Tuple = self._get_next_ind(__magic_name__ ) def lowerCamelCase (self , __magic_name__ , __magic_name__ ) -> None: '''simple docstring''' for ind in self._iterate_buckets(__magic_name__ ): if self._try_set(__magic_name__ , __magic_name__ , __magic_name__ ): break def __setitem__(self , __magic_name__ , __magic_name__ ) -> None: '''simple docstring''' if self._is_full(): self._size_up() self._add_item(__magic_name__ , __magic_name__ ) def __delitem__(self , __magic_name__ ) -> None: '''simple docstring''' for ind in self._iterate_buckets(__magic_name__ ): snake_case_ : Optional[Any] = self._buckets[ind] if item is None: raise KeyError(__magic_name__ ) if item is _deleted: continue if item.key == key: snake_case_ : Union[str, Any] = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__(self , __magic_name__ ) -> VAL: '''simple docstring''' for ind in self._iterate_buckets(__magic_name__ ): snake_case_ : List[Any] = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(__magic_name__ ) def __len__(self ) -> int: '''simple docstring''' return self._len def __iter__(self ) -> Iterator[KEY]: '''simple docstring''' yield from (item.key for item in self._buckets if item) def __repr__(self ) -> str: '''simple docstring''' snake_case_ : List[str] = ''' ,'''.join( F'''{item.key}: {item.val}''' for item in self._buckets if item ) return F'''HashMap({val_string})'''
279
1
"""simple docstring""" from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Tuple = { """microsoft/xprophetnet-large-wiki100-cased""": ( """https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json""" ), } class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ ='xlm-prophetnet' lowerCamelCase__ =['past_key_values'] lowerCamelCase__ ={ 'num_attention_heads': 'num_encoder_attention_heads', } def __init__(self , a_ = 0.1 , a_ = "gelu" , a_ = 3_05_22 , a_ = 10_24 , a_ = 40_96 , a_ = 12 , a_ = 16 , a_ = 40_96 , a_ = 12 , a_ = 16 , a_ = 0.1 , a_ = 0.1 , a_ = 5_12 , a_ = 0.02 , a_ = True , a_ = True , a_ = 0 , a_ = 2 , a_ = 32 , a_ = 1_28 , a_ = False , a_ = 0.0 , a_ = True , a_ = 0 , a_ = 1 , a_ = 2 , **a_ , ): '''simple docstring''' __snake_case : str = vocab_size __snake_case : Optional[int] = hidden_size __snake_case : List[str] = encoder_ffn_dim __snake_case : List[str] = num_encoder_layers __snake_case : Dict = num_encoder_attention_heads __snake_case : Dict = decoder_ffn_dim __snake_case : List[Any] = num_decoder_layers __snake_case : Optional[int] = num_decoder_attention_heads __snake_case : List[str] = max_position_embeddings __snake_case : Any = init_std # Normal(0, this parameter) __snake_case : str = activation_function # parameters for xlmprophetnet __snake_case : Union[str, Any] = ngram __snake_case : Union[str, Any] = num_buckets __snake_case : Optional[int] = relative_max_distance __snake_case : List[Any] = disable_ngram_loss __snake_case : int = eps # 3 Types of Dropout __snake_case : List[str] = attention_dropout __snake_case : Union[str, Any] = activation_dropout __snake_case : Optional[int] = dropout __snake_case : List[str] = use_cache super().__init__( pad_token_id=a_ , bos_token_id=a_ , eos_token_id=a_ , is_encoder_decoder=a_ , add_cross_attention=a_ , decoder_start_token_id=a_ , **a_ , ) @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and''' ''' `num_decoder_layers`.''' )
102
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __lowerCamelCase : str = random.Random() if is_torch_available(): import torch def A_ ( _lowerCAmelCase , _lowerCAmelCase=1.0 , _lowerCAmelCase=None , _lowerCAmelCase=None ) -> Optional[Any]: if rng is None: UpperCamelCase : Optional[int] = global_rng UpperCamelCase : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ): '''simple docstring''' UpperCamelCase : Tuple = parent UpperCamelCase : List[Any] = batch_size UpperCamelCase : List[Any] = min_seq_length UpperCamelCase : List[str] = max_seq_length UpperCamelCase : int = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCamelCase : Union[str, Any] = feature_size UpperCamelCase : List[str] = padding_value UpperCamelCase : Optional[Any] = sampling_rate UpperCamelCase : List[str] = return_attention_mask UpperCamelCase : List[Any] = do_normalize def __UpperCamelCase( self ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __UpperCamelCase( self , A_=False , A_=False ): '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: UpperCamelCase : List[str] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size UpperCamelCase : Dict = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase : Union[str, Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :Optional[Any] = ASTFeatureExtractor def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = ASTFeatureExtractionTester(self ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input UpperCamelCase : Dict = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values UpperCamelCase : Union[str, Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched UpperCamelCase : Any = feat_extract(A_ , padding=A_ , return_tensors="np" ).input_values UpperCamelCase : Any = feat_extract(A_ , padding=A_ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCamelCase : Dict = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCamelCase : int = np.asarray(A_ ) UpperCamelCase : Any = feat_extract(A_ , return_tensors="np" ).input_values UpperCamelCase : List[str] = feat_extract(A_ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch UpperCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase : int = np.random.rand(100 ).astype(np.floataa ) UpperCamelCase : str = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCamelCase : List[Any] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) UpperCamelCase : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __UpperCamelCase( self , A_ ): '''simple docstring''' from datasets import load_dataset UpperCamelCase : Dict = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech UpperCamelCase : Any = ds.sort("id" ).select(range(A_ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = torch.tensor( [-0.98_94, -1.27_76, -0.90_66, -1.27_76, -0.93_49, -1.26_09, -1.03_86, -1.27_76, -1.15_61, -1.27_76, -1.20_52, -1.27_23, -1.21_90, -1.21_32, -1.27_76, -1.11_33, -1.19_53, -1.13_43, -1.15_84, -1.22_03, -1.17_70, -1.24_74, -1.23_81, -1.19_36, -0.92_70, -0.83_17, -0.80_49, -0.77_06, -0.75_65, -0.78_69] ) # fmt: on UpperCamelCase : List[Any] = self._load_datasamples(1 ) UpperCamelCase : Tuple = ASTFeatureExtractor() UpperCamelCase : str = feature_extractor(A_ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 1024, 128) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , A_ , atol=1e-4 ) )
52
0
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowercase = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowercase = direct_transformers_import(PATH_TO_TRANSFORMERS) lowercase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowercase = re.compile(R"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowercase = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCamelCase_ ( UpperCamelCase__ : str ): '''simple docstring''' UpperCamelCase__ = None # source code of `config_class` UpperCamelCase__ = inspect.getsource(UpperCamelCase__ ) UpperCamelCase__ = _re_checkpoint.findall(UpperCamelCase__ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): UpperCamelCase__ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link UpperCamelCase__ = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: UpperCamelCase__ = ckpt_name break return checkpoint def lowerCamelCase_ ( ): '''simple docstring''' UpperCamelCase__ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue UpperCamelCase__ = get_checkpoint_from_config_class(UpperCamelCase__ ) UpperCamelCase__ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: UpperCamelCase__ = '''\n'''.join(sorted(UpperCamelCase__ ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys lowercase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
35
1
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class __lowerCAmelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=9_9 , lowerCAmelCase__=3_2 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__=True , lowerCAmelCase__=5_1_2 , lowerCAmelCase__=1_6 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ) -> Any: '''simple docstring''' a__ : str =parent a__ : Dict =batch_size a__ : List[str] =seq_length a__ : Any =is_training a__ : Tuple =use_input_mask a__ : List[str] =use_token_type_ids a__ : Union[str, Any] =use_labels a__ : Optional[int] =vocab_size a__ : int =hidden_size a__ : int =num_hidden_layers a__ : List[Any] =num_attention_heads a__ : str =intermediate_multiple_size a__ : List[str] =hidden_act a__ : Optional[int] =hidden_dropout a__ : List[str] =attention_dropout a__ : int =weight_tying a__ : Optional[Any] =max_position_embeddings a__ : Any =type_vocab_size a__ : Optional[int] =type_sequence_label_size a__ : Optional[Any] =initializer_range a__ : Dict =num_labels a__ : List[str] =num_choices a__ : Union[str, Any] =scope def _lowercase ( self ) -> Dict: '''simple docstring''' a__ : int =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : Optional[int] =None if self.use_input_mask: a__ : List[str] =random_attention_mask([self.batch_size, self.seq_length] ) a__ : Dict =None if self.use_labels: a__ : str =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ : Any =self.get_config() return config, input_ids, input_mask, token_labels def _lowercase ( self ) -> Dict: '''simple docstring''' return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ , a__ , a__ , a__ : Tuple =self.prepare_config_and_inputs() a__ : List[str] =True return config, input_ids, input_mask, token_labels def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: '''simple docstring''' a__ : Any =GPTNeoXJapaneseModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a__ : List[Any] =model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) a__ : Union[str, Any] =model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: '''simple docstring''' a__ : Optional[int] =True a__ : Dict =GPTNeoXJapaneseModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a__ : Optional[int] =model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' a__ : Optional[int] =GPTNeoXJapaneseForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a__ : Optional[Any] =model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' a__ : int =True a__ : str =GPTNeoXJapaneseForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() # first forward pass a__ : Any =model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) a__ : List[str] =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids a__ : Tuple =ids_tensor((self.batch_size, 3) , config.vocab_size ) a__ : List[Any] =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and a__ : List[str] =torch.cat([input_ids, next_tokens] , dim=-1 ) a__ : List[str] =torch.cat([input_mask, next_mask] , dim=-1 ) a__ : int =model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) a__ : Dict =output_from_no_past["hidden_states"][0] a__ : Any =model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] # select random slice a__ : List[str] =ids_tensor((1,) , output_from_past.shape[-1] ).item() a__ : List[Any] =output_from_no_past[:, -3:, random_slice_idx].detach() a__ : Any =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def _lowercase ( self ) -> List[str]: '''simple docstring''' a__ : List[Any] =self.prepare_config_and_inputs() a__ , a__ , a__ , a__ : int =config_and_inputs a__ : Optional[Any] ={"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): _lowercase : Optional[int] = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () _lowercase : List[str] = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () _lowercase : Optional[int] = ( {"""feature-extraction""": GPTNeoXJapaneseModel, """text-generation""": GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) _lowercase : int = False _lowercase : Optional[Any] = False _lowercase : Tuple = False _lowercase : int = False def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : Optional[Any] =GPTNeoXJapaneseModelTester(self ) a__ : Union[str, Any] =ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=3_7 ) def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ) -> List[Any]: '''simple docstring''' a__ , a__ , a__ , a__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> Any: '''simple docstring''' a__ , a__ , a__ , a__ : List[str] =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> List[str]: '''simple docstring''' a__ , a__ , a__ , a__ : Optional[int] =self.model_tester.prepare_config_and_inputs_for_decoder() a__ : str =None self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ , a__ , a__ , a__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' a__ : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCAmelCase__ ) @slow def _lowercase ( self ) -> int: '''simple docstring''' a__ : Tuple ="abeja/gpt-neox-japanese-2.7b" a__ : Any =["データサイエンティストとは、", "100年後に必要とされる会社は、", "フルリモートの環境で働くために必要なことは、", "国境の長いトンネルを抜けると", "美味しい日本食といえば、"] a__ : Dict =[ "データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。", "100年後に必要とされる会社は、「人」が中心の会社です。", "フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。", "国境の長いトンネルを抜けると、そこは雪国だった。", "美味しい日本食といえば、やっぱりお寿司ですよね。", ] a__ : int =GPTNeoXJapaneseTokenizer.from_pretrained(lowerCAmelCase__ ) a__ : Dict =GPTNeoXJapaneseForCausalLM.from_pretrained(lowerCAmelCase__ ) a__ : List[str] =[] for prompt in prompts: a__ : List[str] =tokenizer(lowerCAmelCase__ , return_tensors="pt" ).input_ids a__ : int =model.generate(lowerCAmelCase__ , max_length=5_0 ) a__ : Dict =tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ )
95
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError('''Length must be a positive.''' ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError('''Length must be a positive.''' ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
101
0
'''simple docstring''' import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging UpperCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE( __lowercase ) -> Tuple: A: int = r'\w+[.]\d+' A: int = re.findall(_A , _A ) for pat in pats: A: Optional[Any] = key.replace(_A , '''_'''.join(pat.split('''.''' ) ) ) return key def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> List[Any]: A: Tuple = pt_tuple_key[:-1] + ('scale',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A: str = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A: Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A: int = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A: Optional[int] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A: Any = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A: Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A: Optional[int] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A: Union[str, Any] = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A: Tuple = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase=4_2 ) -> Optional[Any]: A: Optional[Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A: List[Any] = flax_model.init_weights(PRNGKey(_A ) ) A: List[Any] = flatten_dict(_A ) A: List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A: Union[str, Any] = rename_key(_A ) A: List[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters A: Tuple = rename_key_and_reshape_tensor(_A , _A , _A ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A: Dict = jnp.asarray(_A ) return unflatten_dict(_A )
353
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() UpperCamelCase = logging.get_logger('''transformers.models.encodec''') UpperCamelCase = { '''quantizer.vq.layers.*._codebook.inited''': '''quantizer.layers.*.codebook.inited''', '''quantizer.vq.layers.*._codebook.cluster_size''': '''quantizer.layers.*.codebook.cluster_size''', '''quantizer.vq.layers.*._codebook.embed''': '''quantizer.layers.*.codebook.embed''', '''quantizer.vq.layers.*._codebook.embed_avg''': '''quantizer.layers.*.codebook.embed_avg''', } UpperCamelCase = { '''encoder.model.0.conv.conv''': '''encoder.layers.0.conv''', '''encoder.model.1.block.1.conv.conv''': '''encoder.layers.1.block.1.conv''', '''encoder.model.1.block.3.conv.conv''': '''encoder.layers.1.block.3.conv''', '''encoder.model.1.shortcut.conv.conv''': '''encoder.layers.1.shortcut.conv''', '''encoder.model.3.conv.conv''': '''encoder.layers.3.conv''', '''encoder.model.4.block.1.conv.conv''': '''encoder.layers.4.block.1.conv''', '''encoder.model.4.block.3.conv.conv''': '''encoder.layers.4.block.3.conv''', '''encoder.model.4.shortcut.conv.conv''': '''encoder.layers.4.shortcut.conv''', '''encoder.model.6.conv.conv''': '''encoder.layers.6.conv''', '''encoder.model.7.block.1.conv.conv''': '''encoder.layers.7.block.1.conv''', '''encoder.model.7.block.3.conv.conv''': '''encoder.layers.7.block.3.conv''', '''encoder.model.7.shortcut.conv.conv''': '''encoder.layers.7.shortcut.conv''', '''encoder.model.9.conv.conv''': '''encoder.layers.9.conv''', '''encoder.model.10.block.1.conv.conv''': '''encoder.layers.10.block.1.conv''', '''encoder.model.10.block.3.conv.conv''': '''encoder.layers.10.block.3.conv''', '''encoder.model.10.shortcut.conv.conv''': '''encoder.layers.10.shortcut.conv''', '''encoder.model.12.conv.conv''': '''encoder.layers.12.conv''', '''encoder.model.13.lstm''': '''encoder.layers.13.lstm''', '''encoder.model.15.conv.conv''': '''encoder.layers.15.conv''', } UpperCamelCase = { '''encoder.model.0.conv.norm''': '''encoder.layers.0.norm''', '''encoder.model.1.block.1.conv.norm''': '''encoder.layers.1.block.1.norm''', '''encoder.model.1.block.3.conv.norm''': '''encoder.layers.1.block.3.norm''', '''encoder.model.1.shortcut.conv.norm''': '''encoder.layers.1.shortcut.norm''', '''encoder.model.3.conv.norm''': '''encoder.layers.3.norm''', '''encoder.model.4.block.1.conv.norm''': '''encoder.layers.4.block.1.norm''', '''encoder.model.4.block.3.conv.norm''': '''encoder.layers.4.block.3.norm''', '''encoder.model.4.shortcut.conv.norm''': '''encoder.layers.4.shortcut.norm''', '''encoder.model.6.conv.norm''': '''encoder.layers.6.norm''', '''encoder.model.7.block.1.conv.norm''': '''encoder.layers.7.block.1.norm''', '''encoder.model.7.block.3.conv.norm''': '''encoder.layers.7.block.3.norm''', '''encoder.model.7.shortcut.conv.norm''': '''encoder.layers.7.shortcut.norm''', '''encoder.model.9.conv.norm''': '''encoder.layers.9.norm''', '''encoder.model.10.block.1.conv.norm''': '''encoder.layers.10.block.1.norm''', '''encoder.model.10.block.3.conv.norm''': '''encoder.layers.10.block.3.norm''', '''encoder.model.10.shortcut.conv.norm''': '''encoder.layers.10.shortcut.norm''', '''encoder.model.12.conv.norm''': '''encoder.layers.12.norm''', '''encoder.model.15.conv.norm''': '''encoder.layers.15.norm''', } UpperCamelCase = { '''decoder.model.0.conv.conv''': '''decoder.layers.0.conv''', '''decoder.model.1.lstm''': '''decoder.layers.1.lstm''', '''decoder.model.3.convtr.convtr''': '''decoder.layers.3.conv''', '''decoder.model.4.block.1.conv.conv''': '''decoder.layers.4.block.1.conv''', '''decoder.model.4.block.3.conv.conv''': '''decoder.layers.4.block.3.conv''', '''decoder.model.4.shortcut.conv.conv''': '''decoder.layers.4.shortcut.conv''', '''decoder.model.6.convtr.convtr''': '''decoder.layers.6.conv''', '''decoder.model.7.block.1.conv.conv''': '''decoder.layers.7.block.1.conv''', '''decoder.model.7.block.3.conv.conv''': '''decoder.layers.7.block.3.conv''', '''decoder.model.7.shortcut.conv.conv''': '''decoder.layers.7.shortcut.conv''', '''decoder.model.9.convtr.convtr''': '''decoder.layers.9.conv''', '''decoder.model.10.block.1.conv.conv''': '''decoder.layers.10.block.1.conv''', '''decoder.model.10.block.3.conv.conv''': '''decoder.layers.10.block.3.conv''', '''decoder.model.10.shortcut.conv.conv''': '''decoder.layers.10.shortcut.conv''', '''decoder.model.12.convtr.convtr''': '''decoder.layers.12.conv''', '''decoder.model.13.block.1.conv.conv''': '''decoder.layers.13.block.1.conv''', '''decoder.model.13.block.3.conv.conv''': '''decoder.layers.13.block.3.conv''', '''decoder.model.13.shortcut.conv.conv''': '''decoder.layers.13.shortcut.conv''', '''decoder.model.15.conv.conv''': '''decoder.layers.15.conv''', } UpperCamelCase = { '''decoder.model.0.conv.norm''': '''decoder.layers.0.norm''', '''decoder.model.3.convtr.norm''': '''decoder.layers.3.norm''', '''decoder.model.4.block.1.conv.norm''': '''decoder.layers.4.block.1.norm''', '''decoder.model.4.block.3.conv.norm''': '''decoder.layers.4.block.3.norm''', '''decoder.model.4.shortcut.conv.norm''': '''decoder.layers.4.shortcut.norm''', '''decoder.model.6.convtr.norm''': '''decoder.layers.6.norm''', '''decoder.model.7.block.1.conv.norm''': '''decoder.layers.7.block.1.norm''', '''decoder.model.7.block.3.conv.norm''': '''decoder.layers.7.block.3.norm''', '''decoder.model.7.shortcut.conv.norm''': '''decoder.layers.7.shortcut.norm''', '''decoder.model.9.convtr.norm''': '''decoder.layers.9.norm''', '''decoder.model.10.block.1.conv.norm''': '''decoder.layers.10.block.1.norm''', '''decoder.model.10.block.3.conv.norm''': '''decoder.layers.10.block.3.norm''', '''decoder.model.10.shortcut.conv.norm''': '''decoder.layers.10.shortcut.norm''', '''decoder.model.12.convtr.norm''': '''decoder.layers.12.norm''', '''decoder.model.13.block.1.conv.norm''': '''decoder.layers.13.block.1.norm''', '''decoder.model.13.block.3.conv.norm''': '''decoder.layers.13.block.3.norm''', '''decoder.model.13.shortcut.conv.norm''': '''decoder.layers.13.shortcut.norm''', '''decoder.model.15.conv.norm''': '''decoder.layers.15.norm''', } UpperCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } UpperCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } UpperCamelCase = [] UpperCamelCase = [] def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Dict: for attribute in key.split('''.''' ): A: Union[str, Any] = getattr(__lowercase , __lowercase ) if weight_type is not None: A: Tuple = getattr(__lowercase , __lowercase ).shape else: A: str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": A: Dict = value elif weight_type == "weight_g": A: Tuple = value elif weight_type == "weight_v": A: Any = value elif weight_type == "bias": A: str = value elif weight_type == "running_mean": A: List[Any] = value elif weight_type == "running_var": A: Dict = value elif weight_type == "num_batches_tracked": A: List[str] = value elif weight_type == "weight_ih_l0": A: Dict = value elif weight_type == "weight_hh_l0": A: Optional[int] = value elif weight_type == "bias_ih_l0": A: List[Any] = value elif weight_type == "bias_hh_l0": A: str = value elif weight_type == "weight_ih_l1": A: Optional[int] = value elif weight_type == "weight_hh_l1": A: int = value elif weight_type == "bias_ih_l1": A: Optional[Any] = value elif weight_type == "bias_hh_l1": A: str = value else: A: Optional[int] = value logger.info(F"""{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.""" ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Optional[Any]: for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: A , A: Any = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> Tuple: A: Any = [] if model_name == "encodec_24khz" or "encodec_32khz": A: List[str] = MAPPING_24K elif model_name == "encodec_48khz": A: List[Any] = MAPPING_48K else: raise ValueError(F"""Unsupported model: {model_name}""" ) for name, value in orig_dict.items(): if should_ignore(__lowercase , __lowercase ): logger.info(F"""{name} was ignored""" ) continue A: Optional[int] = False for key, mapped_key in MAPPING.items(): if "*" in key: A , A: Optional[int] = key.split('''.*.''' ) if prefix in name and suffix in name: A: str = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue A: Optional[Any] = True if "*" in mapped_key: A: Any = name.split(__lowercase )[0].split('''.''' )[-2] A: Tuple = mapped_key.replace('''*''' , __lowercase ) if "weight_g" in name: A: str = '''weight_g''' elif "weight_v" in name: A: List[Any] = '''weight_v''' elif "weight_ih_l0" in name: A: Dict = '''weight_ih_l0''' elif "weight_hh_l0" in name: A: int = '''weight_hh_l0''' elif "bias_ih_l0" in name: A: Union[str, Any] = '''bias_ih_l0''' elif "bias_hh_l0" in name: A: Tuple = '''bias_hh_l0''' elif "weight_ih_l1" in name: A: int = '''weight_ih_l1''' elif "weight_hh_l1" in name: A: Optional[Any] = '''weight_hh_l1''' elif "bias_ih_l1" in name: A: Dict = '''bias_ih_l1''' elif "bias_hh_l1" in name: A: str = '''bias_hh_l1''' elif "bias" in name: A: Union[str, Any] = '''bias''' elif "weight" in name: A: Dict = '''weight''' elif "running_mean" in name: A: Tuple = '''running_mean''' elif "running_var" in name: A: Any = '''running_var''' elif "num_batches_tracked" in name: A: str = '''num_batches_tracked''' else: A: Tuple = None set_recursively(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) continue if not is_used: unused_weights.append(__lowercase ) logger.warning(F"""Unused weights: {unused_weights}""" ) @torch.no_grad() def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , ) -> Dict: if config_path is not None: A: Tuple = EncodecConfig.from_pretrained(__lowercase ) else: A: Union[str, Any] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": A: Union[str, Any] = [8, 5, 4, 4] A: Dict = [2.2] A: List[Any] = 6_4 A: Optional[Any] = 3_2_0_0_0 A: List[Any] = 2_0_4_8 A: Optional[Any] = False A: int = False A: Union[str, Any] = False elif model_name == "encodec_48khz": A: Optional[int] = [8, 5, 4, 2] A: List[Any] = [3.0, 6.0, 1_2.0, 2_4.0] A: List[Any] = 4_8_0_0_0 A: int = 2 A: List[Any] = False A: Any = '''time_group_norm''' A: Optional[Any] = True A: Any = 1.0 A: Any = 0.0_1 else: raise ValueError(F"""Unknown model name: {model_name}""" ) A: str = EncodecModel(__lowercase ) A: Optional[Any] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(__lowercase ) A: Union[str, Any] = torch.load(__lowercase ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights A: Optional[int] = original_checkpoint['''best_state'''] recursively_load_weights(__lowercase , __lowercase , __lowercase ) model.save_pretrained(__lowercase ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(__lowercase ) model.push_to_hub(__lowercase ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model''', default='''encodec_24khz''', type=str, help='''The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) UpperCamelCase = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
334
0
from math import sqrt def __A ( __lowerCamelCase = 100_0000 ) -> int: a = 0 a = 0 a = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F'{solution() = }')
228
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() __UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) __UpperCamelCase : Any = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } __UpperCamelCase : Dict = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def __A ( __lowerCamelCase ) -> List[str]: a = {} with open(__lowerCamelCase , """r""" ) as file: for line_number, line in enumerate(__lowerCamelCase ): a = line.strip() if line: a = line.split() a = line_number a = words[0] a = value return result def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Dict: for attribute in key.split(""".""" ): a = getattr(__lowerCamelCase , __lowerCamelCase ) a = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(__lowerCamelCase ): a = PARAM_MAPPING[full_name.split(""".""" )[-1]] a = """param""" if weight_type is not None and weight_type != "param": a = getattr(__lowerCamelCase , __lowerCamelCase ).shape elif weight_type is not None and weight_type == "param": a = hf_pointer for attribute in hf_param_name.split(""".""" ): a = getattr(__lowerCamelCase , __lowerCamelCase ) a = shape_pointer.shape # let's reduce dimension a = value[0] else: a = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": a = value elif weight_type == "weight_g": a = value elif weight_type == "weight_v": a = value elif weight_type == "bias": a = value elif weight_type == "param": for attribute in hf_param_name.split(""".""" ): a = getattr(__lowerCamelCase , __lowerCamelCase ) a = value else: a = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: a = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(__lowerCamelCase ): a = PARAM_MAPPING[full_name.split(""".""" )[-1]] a = """param""" if weight_type is not None and weight_type != "param": a = """.""".join([key, weight_type] ) elif weight_type is not None and weight_type == "param": a = """.""".join([key, hf_param_name] ) else: a = key a = value if """lm_head""" in full_key else value[0] __UpperCamelCase : List[Any] = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None ) -> Optional[Any]: a = False for key, mapped_key in MAPPING.items(): a = """wav2vec2.""" + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: a = True if "*" in mapped_key: a = name.split(__lowerCamelCase )[0].split(""".""" )[-2] a = mapped_key.replace("""*""" , __lowerCamelCase ) if "weight_g" in name: a = """weight_g""" elif "weight_v" in name: a = """weight_v""" elif "bias" in name: a = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj a = """weight""" else: a = None if hf_dict is not None: rename_dict(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: set_recursively(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return is_used return is_used def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> List[Any]: a = [] a = fairseq_model.state_dict() a = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): a = False if "conv_layers" in name: load_conv_layer( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , hf_model.config.feat_extract_norm == """group""" , ) a = True else: a = load_wavaveca_layer(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not is_used: unused_weights.append(__lowerCamelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: a = full_name.split("""conv_layers.""" )[-1] a = name.split(""".""" ) a = int(items[0] ) a = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) a = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) a = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) a = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) a = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__lowerCamelCase ) @torch.no_grad() def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=True , __lowerCamelCase=False ) -> List[Any]: if config_path is not None: a = WavaVecaConfig.from_pretrained(__lowerCamelCase ) else: a = WavaVecaConfig() if is_seq_class: a = read_txt_into_dict(__lowerCamelCase ) a = idalabel a = WavaVecaForSequenceClassification(__lowerCamelCase ) a = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__lowerCamelCase , return_attention_mask=__lowerCamelCase , ) feature_extractor.save_pretrained(__lowerCamelCase ) elif is_finetuned: if dict_path: a = Dictionary.load(__lowerCamelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq a = target_dict.pad_index a = target_dict.bos_index a = target_dict.eos_index a = len(target_dict.symbols ) a = os.path.join(__lowerCamelCase , """vocab.json""" ) if not os.path.isdir(__lowerCamelCase ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(__lowerCamelCase ) ) return os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) a = target_dict.indices # fairseq has the <pad> and <s> switched a = 0 a = 1 with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(__lowerCamelCase , __lowerCamelCase ) a = WavaVecaCTCTokenizer( __lowerCamelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=__lowerCamelCase , ) a = True if config.feat_extract_norm == """layer""" else False a = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__lowerCamelCase , return_attention_mask=__lowerCamelCase , ) a = WavaVecaProcessor(feature_extractor=__lowerCamelCase , tokenizer=__lowerCamelCase ) processor.save_pretrained(__lowerCamelCase ) a = WavaVecaForCTC(__lowerCamelCase ) else: a = WavaVecaForPreTraining(__lowerCamelCase ) if is_finetuned or is_seq_class: a , a , a = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) else: a = argparse.Namespace(task="""audio_pretraining""" ) a = fairseq.tasks.setup_task(__lowerCamelCase ) a , a , a = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=__lowerCamelCase ) a = model[0].eval() recursively_load_weights(__lowerCamelCase , __lowerCamelCase , not is_finetuned ) hf_wavavec.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": __UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) __UpperCamelCase : Union[str, Any] = parser.parse_args() __UpperCamelCase : Any = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
228
1
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model _UpperCAmelCase : Tuple = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def A ( lowercase , lowercase , lowercase=None ) -> Optional[int]: '''simple docstring''' if rng is None: UpperCamelCase = random.Random() UpperCamelCase = 1 for dim in shape: total_dims *= dim UpperCamelCase = [] for _ in range(lowercase ): values.append(rng.randint(0 , vocab_size - 1 ) ) UpperCamelCase = np.array(lowercase , dtype=jnp.intaa ).reshape(lowercase ) return output def A ( lowercase , lowercase=None ) -> str: '''simple docstring''' UpperCamelCase = ids_tensor(lowercase , vocab_size=2 , rng=lowercase ) # make sure that at least one token is attended to for each batch UpperCamelCase = 1 return attn_mask @require_flax class lowercase : __lowercase : Any = None __lowercase : List[str] = () def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 UpperCamelCase = 2 UpperCamelCase = inputs['input_ids'].shape[-1] // 2 UpperCamelCase = inputs['input_ids'][:max_batch_size, :sequence_length] UpperCamelCase = jnp.ones_like(A_ ) UpperCamelCase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens UpperCamelCase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` UpperCamelCase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def __UpperCamelCase ( self ) -> Tuple: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 0 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning UpperCamelCase = getattr(A_ , A_ ) UpperCamelCase = pt_model_class(A_ ).eval() UpperCamelCase = load_flax_weights_in_pytorch_model(A_ , flax_model.params ) UpperCamelCase = flax_model.generate(A_ ).sequences UpperCamelCase = pt_model.generate(torch.tensor(A_ , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: UpperCamelCase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> Tuple: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = True UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 2 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 2 UpperCamelCase = 2 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = True UpperCamelCase = max_length UpperCamelCase = 0.8 UpperCamelCase = 10 UpperCamelCase = 0.3 UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = max_length UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = max_length UpperCamelCase = 2 UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = False UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ , attention_mask=A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ , attention_mask=A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = True UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ , attention_mask=A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ , attention_mask=A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = 2 UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(A_ ) UpperCamelCase = model.generate(A_ , attention_mask=A_ ).sequences self.assertEqual(generation_outputs.shape[-1] , A_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(A_ , attention_mask=A_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class lowercase ( unittest.TestCase ): def __UpperCamelCase ( self ) -> Any: """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-bert' ) UpperCamelCase = FlaxAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-bert-flax-only' ) UpperCamelCase = 'Hello world' UpperCamelCase = tokenizer(A_ , return_tensors='np' ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A_ , 'do_samples' ): model.generate(A_ , do_samples=A_ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A_ , 'foo' ): UpperCamelCase = {'foo': 'bar'} model.generate(A_ , **A_ )
110
_UpperCAmelCase : str = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] _UpperCAmelCase : Any = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] _UpperCAmelCase : List[Any] = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] _UpperCAmelCase : Tuple = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] _UpperCAmelCase : Union[str, Any] = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] _UpperCAmelCase : List[str] = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] _UpperCAmelCase : Tuple = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] _UpperCAmelCase : Dict = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
110
1
# Function to print upper half of diamond (pyramid) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' for i in range(0 , lowerCAmelCase__ ): for _ in range(0 , n - i - 1 ): # printing spaces print(''' ''' , end='''''' ) for _ in range(0 , i + 1 ): # printing stars print('''* ''' , end='''''' ) print() def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' for i in range(lowerCAmelCase__ , 0 , -1 ): for _ in range(lowerCAmelCase__ , 0 , -1 ): # printing stars print('''* ''' , end='''''' ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(''' ''' , end='''''' ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if n <= 0: print(''' ... .... nothing printing :(''' ) return floyd(lowerCAmelCase__ ) # upper half reverse_floyd(lowerCAmelCase__ ) # lower half if __name__ == "__main__": print(r"| /\ | |- | |- |--| |\ /| |-") print(r"|/ \| |- |_ |_ |__| | \/ | |_") lowercase__ :Dict = 1 while K: lowercase__ :Optional[Any] = int(input("enter the number and , and see the magic : ")) print() pretty_print(user_number) lowercase__ :Any = int(input("press 0 to exit... and 1 to continue...")) print("Good Bye...")
101
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() lowercase__ :str = 2 class lowercase : def __init__( self ,*, # begin keyword-only arguments A__="<s>" ,A__="<pad>" ,A__="</s>" ,A__="<unk>" ,A__=None ,): lowercase , lowercase , lowercase , lowercase = bos, unk, pad, eos lowercase = [] lowercase = [] lowercase = {} lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(A__) lowercase = len(self.symbols) def __eq__( self ,A__): return self.indices == other.indices def __getitem__( self ,A__): if idx < len(self.symbols): return self.symbols[idx] return self.unk_word def __len__( self): return len(self.symbols) def __contains__( self ,A__): return sym in self.indices @classmethod def A__ ( cls ,A__): lowercase = cls() d.add_from_file(A__) return d def A__ ( self ,A__ ,A__=1 ,A__=False): if word in self.indices and not overwrite: lowercase = self.indices[word] lowercase = self.count[idx] + n return idx else: lowercase = len(self.symbols) lowercase = idx self.symbols.append(A__) self.count.append(A__) return idx def A__ ( self ,A__): return 0 def A__ ( self ,A__): if isinstance(A__ ,A__): try: with open(A__ ,'''r''' ,encoding='''utf-8''') as fd: self.add_from_file(A__) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(A__)) return lowercase = f.readlines() lowercase = self._load_meta(A__) for line in lines[indices_start_line:]: try: lowercase , lowercase = line.rstrip().rsplit(''' ''' ,1) if field == "#fairseq:overwrite": lowercase = True lowercase , lowercase = line.rsplit(''' ''' ,1) else: lowercase = False lowercase = int(A__) lowercase = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(A__)) self.add_symbol(A__ ,n=A__ ,overwrite=A__) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''') def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} lowercase = dict((re.sub(R'''@@$''' , '''''' , lowerCAmelCase__ ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , lowerCAmelCase__ ), v) for k, v in d.items() ) lowercase = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[f'{k}</w>'] lowercase = d[k] # restore return da def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # prep if not os.path.exists(lowerCAmelCase__ ): raise ValueError(f'path {biogpt_checkpoint_path} does not exist!' ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) print(f'Writing results to {pytorch_dump_folder_path}' ) # handle various types of models lowercase = os.path.join(lowerCAmelCase__ , '''checkpoint.pt''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {checkpoint_file} does not exist!' ) lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = chkpt['''cfg''']['''model'''] # dicts lowercase = os.path.join(lowerCAmelCase__ , '''dict.txt''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {dict_file} does not exist!' ) lowercase = Dictionary.load(lowerCAmelCase__ ) lowercase = rewrite_dict_keys(src_dict.indices ) lowercase = len(lowerCAmelCase__ ) lowercase = os.path.join(lowerCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) print(f'Generating {src_vocab_file} of {src_vocab_size} records' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # merges_file (bpecodes) lowercase = os.path.join(lowerCAmelCase__ , '''bpecodes''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {bpecodes_file} does not exist!' ) lowercase = os.path.join(lowerCAmelCase__ , VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) # model config lowercase = os.path.join(lowerCAmelCase__ , '''config.json''' ) lowercase = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.02, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1E-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(f'Generating {biogpt_model_config_file}' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # tokenizer config lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 1024, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(f'Generating {biogpt_tokenizer_config_file}' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # model lowercase = chkpt['''model'''] # remove unneeded keys lowercase = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): lowercase = model_state_dict.pop(lowerCAmelCase__ ) else: lowercase = model_state_dict.pop(lowerCAmelCase__ ) lowercase = BioGptConfig.from_pretrained(lowerCAmelCase__ ) lowercase = BioGptForCausalLM(lowerCAmelCase__ ) # check that it loads ok model_new.load_state_dict(lowerCAmelCase__ ) # save lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) print(f'Generating {pytorch_weights_dump_path}' ) torch.save(lowerCAmelCase__ , lowerCAmelCase__ ) print('''Conversion is done!''' ) if __name__ == "__main__": lowercase__ :Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--biogpt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) lowercase__ :Any = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
101
1
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase__ = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } lowerCAmelCase__ = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } lowerCAmelCase__ = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } lowerCAmelCase__ = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } lowerCAmelCase__ = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } lowerCAmelCase__ = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } lowerCAmelCase__ = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } lowerCAmelCase__ = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } lowerCAmelCase__ = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = VOCAB_FILES_NAMES UpperCAmelCase_ = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = VOCAB_FILES_NAMES UpperCAmelCase_ = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION lowerCAmelCase__ = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) lowerCAmelCase__ = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) lowerCAmelCase__ = r'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n ```\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n ```\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Returns:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(SCREAMING_SNAKE_CASE ) class a_ : '''simple docstring''' def __call__( self : Optional[int] , lowercase__ : List[str] , lowercase__ : Optional[str] = None , lowercase__ : Optional[str] = None , lowercase__ : Union[bool, str] = False , lowercase__ : Union[bool, str] = False , lowercase__ : Optional[int] = None , lowercase__ : Optional[Union[str, TensorType]] = None , lowercase__ : Optional[bool] = None , **lowercase__ : Union[str, Any] , ): '''simple docstring''' if titles is None and texts is None: return super().__call__( lowercase__ , padding=lowercase__ , truncation=lowercase__ , max_length=lowercase__ , return_tensors=lowercase__ , return_attention_mask=lowercase__ , **lowercase__ , ) elif titles is None or texts is None: lowerCAmelCase__ = titles if texts is None else texts return super().__call__( lowercase__ , lowercase__ , padding=lowercase__ , truncation=lowercase__ , max_length=lowercase__ , return_tensors=lowercase__ , return_attention_mask=lowercase__ , **lowercase__ , ) lowerCAmelCase__ = titles if not isinstance(lowercase__ , lowercase__) else [titles] lowerCAmelCase__ = texts if not isinstance(lowercase__ , lowercase__) else [texts] lowerCAmelCase__ = len(lowercase__) lowerCAmelCase__ = questions if not isinstance(lowercase__ , lowercase__) else [questions] * n_passages if len(lowercase__) != len(lowercase__): raise ValueError( F"""There should be as many titles than texts but got {len(lowercase__)} titles and {len(lowercase__)} texts.""") lowerCAmelCase__ = super().__call__(lowercase__ , lowercase__ , padding=lowercase__ , truncation=lowercase__)['input_ids'] lowerCAmelCase__ = super().__call__(lowercase__ , add_special_tokens=lowercase__ , padding=lowercase__ , truncation=lowercase__)['input_ids'] lowerCAmelCase__ = { 'input_ids': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowercase__ , lowercase__) ] } if return_attention_mask is not False: lowerCAmelCase__ = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id) for input_id in input_ids]) lowerCAmelCase__ = attention_mask return self.pad(lowercase__ , padding=lowercase__ , max_length=lowercase__ , return_tensors=lowercase__) def __snake_case ( self : Union[str, Any] , lowercase__ : BatchEncoding , lowercase__ : DPRReaderOutput , lowercase__ : int = 16 , lowercase__ : int = 64 , lowercase__ : int = 4 , ): '''simple docstring''' lowerCAmelCase__ = reader_input['input_ids'] lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = reader_output[:3] lowerCAmelCase__ = len(lowercase__) lowerCAmelCase__ = sorted(range(lowercase__) , reverse=lowercase__ , key=relevance_logits.__getitem__) lowerCAmelCase__ = [] for doc_id in sorted_docs: lowerCAmelCase__ = list(input_ids[doc_id]) # assuming question & title information is at the beginning of the sequence lowerCAmelCase__ = sequence_ids.index(self.sep_token_id , 2) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: lowerCAmelCase__ = sequence_ids.index(self.pad_token_id) else: lowerCAmelCase__ = len(lowercase__) lowerCAmelCase__ = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=lowercase__ , top_spans=lowercase__ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=lowercase__ , start_index=lowercase__ , end_index=lowercase__ , text=self.decode(sequence_ids[start_index : end_index + 1]) , )) if len(lowercase__) >= num_spans: break return nbest_spans_predictions[:num_spans] def __snake_case ( self : Optional[int] , lowercase__ : List[int] , lowercase__ : List[int] , lowercase__ : int , lowercase__ : int , ): '''simple docstring''' lowerCAmelCase__ = [] for start_index, start_score in enumerate(lowercase__): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length]): scores.append(((start_index, start_index + answer_length), start_score + end_score)) lowerCAmelCase__ = sorted(lowercase__ , key=lambda lowercase__: x[1] , reverse=lowercase__) lowerCAmelCase__ = [] for (start_index, end_index), score in scores: if start_index > end_index: raise ValueError(F"""Wrong span indices: [{start_index}:{end_index}]""") lowerCAmelCase__ = end_index - start_index + 1 if length > max_answer_length: raise ValueError(F"""Span is too long: {length} > {max_answer_length}""") if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals): continue chosen_span_intervals.append((start_index, end_index)) if len(lowercase__) == top_spans: break return chosen_span_intervals @add_end_docstrings(SCREAMING_SNAKE_CASE ) class a_ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = VOCAB_FILES_NAMES UpperCAmelCase_ = READER_PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ = READER_PRETRAINED_INIT_CONFIGURATION UpperCAmelCase_ = ['input_ids', 'attention_mask']
119
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'adapter_layer': 'encoder.layers.*.adapter_layer', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', 'pooling_layer.linear': 'projector', 'pooling_layer.projection': 'classifier', } lowerCAmelCase__ = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', 'projector', 'classifier', ] def __lowerCamelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = {} with open(lowerCAmelCase__ , 'r' ) as file: for line_number, line in enumerate(lowerCAmelCase__ ): lowerCAmelCase__ = line.strip() if line: lowerCAmelCase__ = line.split() lowerCAmelCase__ = line_number lowerCAmelCase__ = words[0] lowerCAmelCase__ = value return result def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): for attribute in key.split('.' ): lowerCAmelCase__ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCAmelCase__ ): lowerCAmelCase__ = PARAM_MAPPING[full_name.split('.' )[-1]] lowerCAmelCase__ = 'param' if weight_type is not None and weight_type != "param": lowerCAmelCase__ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ).shape elif weight_type is not None and weight_type == "param": lowerCAmelCase__ = hf_pointer for attribute in hf_param_name.split('.' ): lowerCAmelCase__ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ = shape_pointer.shape # let's reduce dimension lowerCAmelCase__ = value[0] else: lowerCAmelCase__ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowerCAmelCase__ = value elif weight_type == "weight_g": lowerCAmelCase__ = value elif weight_type == "weight_v": lowerCAmelCase__ = value elif weight_type == "bias": lowerCAmelCase__ = value elif weight_type == "param": for attribute in hf_param_name.split('.' ): lowerCAmelCase__ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ = value else: lowerCAmelCase__ = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase__ = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCAmelCase__ ): lowerCAmelCase__ = PARAM_MAPPING[full_name.split('.' )[-1]] lowerCAmelCase__ = 'param' if weight_type is not None and weight_type != "param": lowerCAmelCase__ = '.'.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": lowerCAmelCase__ = '.'.join([key, hf_param_name] ) else: lowerCAmelCase__ = key lowerCAmelCase__ = value if 'lm_head' in full_key else value[0] lowerCAmelCase__ = { 'W_a': 'linear_1.weight', 'W_b': 'linear_2.weight', 'b_a': 'linear_1.bias', 'b_b': 'linear_2.bias', 'ln_W': 'norm.weight', 'ln_b': 'norm.bias', } def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ): lowerCAmelCase__ = False for key, mapped_key in MAPPING.items(): lowerCAmelCase__ = 'wav2vec2.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: lowerCAmelCase__ = True if "*" in mapped_key: lowerCAmelCase__ = name.split(lowerCAmelCase__ )[0].split('.' )[-2] lowerCAmelCase__ = mapped_key.replace('*' , lowerCAmelCase__ ) if "weight_g" in name: lowerCAmelCase__ = 'weight_g' elif "weight_v" in name: lowerCAmelCase__ = 'weight_v' elif "bias" in name: lowerCAmelCase__ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj lowerCAmelCase__ = 'weight' else: lowerCAmelCase__ = None if hf_dict is not None: rename_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: set_recursively(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return is_used return is_used def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase__ = [] lowerCAmelCase__ = fairseq_model.state_dict() lowerCAmelCase__ = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): lowerCAmelCase__ = False if "conv_layers" in name: load_conv_layer( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , hf_model.config.feat_extract_norm == 'group' , ) lowerCAmelCase__ = True else: lowerCAmelCase__ = load_wavaveca_layer(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if not is_used: unused_weights.append(lowerCAmelCase__ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase__ = full_name.split('conv_layers.' )[-1] lowerCAmelCase__ = name.split('.' ) lowerCAmelCase__ = int(items[0] ) lowerCAmelCase__ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowerCAmelCase__ = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowerCAmelCase__ = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) lowerCAmelCase__ = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) lowerCAmelCase__ = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowerCAmelCase__ ) @torch.no_grad() def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=False ): if config_path is not None: lowerCAmelCase__ = WavaVecaConfig.from_pretrained(lowerCAmelCase__ ) else: lowerCAmelCase__ = WavaVecaConfig() if is_seq_class: lowerCAmelCase__ = read_txt_into_dict(lowerCAmelCase__ ) lowerCAmelCase__ = idalabel lowerCAmelCase__ = WavaVecaForSequenceClassification(lowerCAmelCase__ ) lowerCAmelCase__ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , ) feature_extractor.save_pretrained(lowerCAmelCase__ ) elif is_finetuned: if dict_path: lowerCAmelCase__ = Dictionary.load(lowerCAmelCase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCAmelCase__ = target_dict.pad_index lowerCAmelCase__ = target_dict.bos_index lowerCAmelCase__ = target_dict.eos_index lowerCAmelCase__ = len(target_dict.symbols ) lowerCAmelCase__ = os.path.join(lowerCAmelCase__ , 'vocab.json' ) if not os.path.isdir(lowerCAmelCase__ ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(lowerCAmelCase__ ) ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) lowerCAmelCase__ = target_dict.indices # fairseq has the <pad> and <s> switched lowerCAmelCase__ = 0 lowerCAmelCase__ = 1 with open(lowerCAmelCase__ , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ = WavaVecaCTCTokenizer( lowerCAmelCase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=lowerCAmelCase__ , ) lowerCAmelCase__ = True if config.feat_extract_norm == 'layer' else False lowerCAmelCase__ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , ) lowerCAmelCase__ = WavaVecaProcessor(feature_extractor=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) processor.save_pretrained(lowerCAmelCase__ ) lowerCAmelCase__ = WavaVecaForCTC(lowerCAmelCase__ ) else: lowerCAmelCase__ = WavaVecaForPreTraining(lowerCAmelCase__ ) if is_finetuned or is_seq_class: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: lowerCAmelCase__ = argparse.Namespace(task='audio_pretraining' ) lowerCAmelCase__ = fairseq.tasks.setup_task(lowerCAmelCase__ ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowerCAmelCase__ ) lowerCAmelCase__ = model[0].eval() recursively_load_weights(lowerCAmelCase__ , lowerCAmelCase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) parser.add_argument( '--is_seq_class', action='store_true', help='Whether the model to convert is a fine-tuned sequence classification model or not', ) lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
119
1
'''simple docstring''' import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version _lowercase : Optional[int] = logging.getLogger(__name__) require_version("pytorch_lightning>=1.0.4") _lowercase : List[str] = { "base": AutoModel, "sequence-classification": AutoModelForSequenceClassification, "question-answering": AutoModelForQuestionAnswering, "pretraining": AutoModelForPreTraining, "token-classification": AutoModelForTokenClassification, "language-modeling": AutoModelWithLMHead, "summarization": AutoModelForSeqaSeqLM, "translation": AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization _lowercase : List[str] = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } _lowercase : Dict = sorted(arg_to_scheduler.keys()) _lowercase : Union[str, Any] = "{" + ", ".join(arg_to_scheduler_choices) + "}" class __magic_name__ ( pl.LightningModule): def __init__( self : Dict , lowercase_ : argparse.Namespace , lowercase_ : str=None , lowercase_ : List[str]="base" , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Tuple=None , **lowercase_ : List[Any] , ): super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(lowercase_ ) lowercase_ : Dict = 0 lowercase_ : Any = Path(self.hparams.output_dir ) lowercase_ : Union[str, Any] = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: lowercase_ : Any = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({"""num_labels""": num_labels} if num_labels is not None else {}) , cache_dir=lowercase_ , **lowercase_ , ) else: lowercase_ : PretrainedConfig = config lowercase_ : Any = ("""encoder_layerdrop""", """decoder_layerdrop""", """dropout""", """attention_dropout""") for p in extra_model_params: if getattr(self.hparams , lowercase_ , lowercase_ ): assert hasattr(self.config , lowercase_ ), f'''model config doesn\'t have a `{p}` attribute''' setattr(self.config , lowercase_ , getattr(self.hparams , lowercase_ ) ) if tokenizer is None: lowercase_ : str = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=lowercase_ , ) else: lowercase_ : PreTrainedTokenizer = tokenizer lowercase_ : Tuple = MODEL_MODES[mode] if model is None: lowercase_ : Any = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool(""".ckpt""" in self.hparams.model_name_or_path ) , config=self.config , cache_dir=lowercase_ , ) else: lowercase_ : Union[str, Any] = model def SCREAMING_SNAKE_CASE_ ( self : List[str] , *lowercase_ : Dict , **lowercase_ : str ): lowercase_ : str = self.model_type.from_pretrained(*lowercase_ , **lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : str = arg_to_scheduler[self.hparams.lr_scheduler] lowercase_ : Any = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps() ) lowercase_ : List[str] = {"""scheduler""": scheduler, """interval""": """step""", """frequency""": 1} return scheduler def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Any = self.model lowercase_ : Dict = ["""bias""", """LayerNorm.weight"""] lowercase_ : int = [ { """params""": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters """weight_decay""": self.hparams.weight_decay, }, { """params""": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] if self.hparams.adafactor: lowercase_ : Optional[int] = Adafactor( lowercase_ , lr=self.hparams.learning_rate , scale_parameter=lowercase_ , relative_step=lowercase_ ) else: lowercase_ : Dict = AdamW( lowercase_ , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon ) lowercase_ : List[str] = optimizer lowercase_ : List[str] = self.get_lr_scheduler() return [optimizer], [scheduler] def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : Optional[Any] , lowercase_ : int ): return self.validation_step(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Union[str, Any] ): return self.validation_end(lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Union[str, Any] = max(1 , self.hparams.gpus ) # TODO: consider num_tpu_cores lowercase_ : List[str] = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def SCREAMING_SNAKE_CASE_ ( self : Any , lowercase_ : Optional[Any] ): if stage == "test": lowercase_ : Any = len(self.test_dataloader().dataset ) else: lowercase_ : List[Any] = self.get_dataloader("""train""" , self.hparams.train_batch_size , shuffle=lowercase_ ) lowercase_ : Union[str, Any] = len(self.train_dataloader().dataset ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , lowercase_ : str , lowercase_ : int , lowercase_ : bool = False ): raise NotImplementedError("""You must implement this for your task""" ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): return self.train_loader def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): return self.get_dataloader("""dev""" , self.hparams.eval_batch_size , shuffle=lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): return self.get_dataloader("""test""" , self.hparams.eval_batch_size , shuffle=lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[str] , lowercase_ : List[Any] ): return os.path.join( self.hparams.data_dir , """cached_{}_{}_{}""".format( lowercase_ , list(filter(lowercase_ , self.hparams.model_name_or_path.split("""/""" ) ) ).pop() , str(self.hparams.max_seq_length ) , ) , ) @pl.utilities.rank_zero_only def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Dict[str, Any] ): lowercase_ : List[Any] = self.output_dir.joinpath("""best_tfmr""" ) lowercase_ : Dict = self.step_count self.model.save_pretrained(lowercase_ ) self.tokenizer.save_pretrained(lowercase_ ) @staticmethod def SCREAMING_SNAKE_CASE_ ( lowercase_ : Optional[int] , lowercase_ : Dict ): parser.add_argument( """--model_name_or_path""" , default=lowercase_ , type=lowercase_ , required=lowercase_ , help="""Path to pretrained model or model identifier from huggingface.co/models""" , ) parser.add_argument( """--config_name""" , default="""""" , type=lowercase_ , help="""Pretrained config name or path if not the same as model_name""" ) parser.add_argument( """--tokenizer_name""" , default=lowercase_ , type=lowercase_ , help="""Pretrained tokenizer name or path if not the same as model_name""" , ) parser.add_argument( """--cache_dir""" , default=str(Path(lowercase_ ).parent / """test_run""" / """cache""" ) , type=lowercase_ , help="""Where do you want to store the pre-trained models downloaded from huggingface.co""" , ) parser.add_argument( """--encoder_layerdrop""" , type=lowercase_ , help="""Encoder layer dropout probability (Optional). Goes into model.config""" , ) parser.add_argument( """--decoder_layerdrop""" , type=lowercase_ , help="""Decoder layer dropout probability (Optional). Goes into model.config""" , ) parser.add_argument( """--dropout""" , type=lowercase_ , help="""Dropout probability (Optional). Goes into model.config""" , ) parser.add_argument( """--attention_dropout""" , type=lowercase_ , help="""Attention dropout probability (Optional). Goes into model.config""" , ) parser.add_argument("""--learning_rate""" , default=5E-5 , type=lowercase_ , help="""The initial learning rate for Adam.""" ) parser.add_argument( """--lr_scheduler""" , default="""linear""" , choices=lowercase_ , metavar=lowercase_ , type=lowercase_ , help="""Learning rate scheduler""" , ) parser.add_argument("""--weight_decay""" , default=0.0 , type=lowercase_ , help="""Weight decay if we apply some.""" ) parser.add_argument("""--adam_epsilon""" , default=1E-8 , type=lowercase_ , help="""Epsilon for Adam optimizer.""" ) parser.add_argument("""--warmup_steps""" , default=0 , type=lowercase_ , help="""Linear warmup over warmup_steps.""" ) parser.add_argument("""--num_workers""" , default=4 , type=lowercase_ , help="""kwarg passed to DataLoader""" ) parser.add_argument("""--num_train_epochs""" , dest="""max_epochs""" , default=3 , type=lowercase_ ) parser.add_argument("""--train_batch_size""" , default=32 , type=lowercase_ ) parser.add_argument("""--eval_batch_size""" , default=32 , type=lowercase_ ) parser.add_argument("""--adafactor""" , action="""store_true""" ) class __magic_name__ ( pl.Callback): def SCREAMING_SNAKE_CASE_ ( self : str , lowercase_ : List[Any] , lowercase_ : List[str] ): if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class __magic_name__ ( pl.Callback): def SCREAMING_SNAKE_CASE_ ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : int ): # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(lowercase_ ) class __magic_name__ ( pl.Callback): def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , lowercase_ : str , lowercase_ : Any ): lowercase_ : str = trainer.lr_schedulers[0]["""scheduler"""] lowercase_ : Union[str, Any] = {f'''lr_group_{i}''': lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[str] , lowercase_ : pl.Trainer , lowercase_ : pl.LightningModule ): rank_zero_info("""***** Validation results *****""" ) lowercase_ : List[Any] = trainer.callback_metrics # Log results for key in sorted(lowercase_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(lowercase_ , str(metrics[key] ) ) ) def SCREAMING_SNAKE_CASE_ ( self : str , lowercase_ : pl.Trainer , lowercase_ : pl.LightningModule ): rank_zero_info("""***** Test results *****""" ) lowercase_ : Tuple = trainer.callback_metrics # Log and save results to file lowercase_ : str = os.path.join(pl_module.hparams.output_dir , """test_results.txt""" ) with open(lowercase_ , """w""" ) as writer: for key in sorted(lowercase_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(lowercase_ , str(metrics[key] ) ) ) writer.write("""{} = {}\n""".format(lowercase_ , str(metrics[key] ) ) ) def lowerCamelCase ( UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict ) -> None: # To allow all pl args uncomment the following line # parser = pl.Trainer.add_argparse_args(parser) parser.add_argument( """--output_dir""" , default=str(Path(UpperCAmelCase__ ).parent / """test_run""" / """model_checkpoints""" ) , type=UpperCAmelCase__ , help="""The output directory where the model predictions and checkpoints will be written.""" , ) parser.add_argument( """--fp16""" , action="""store_true""" , help="""Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit""" , ) parser.add_argument( """--fp16_opt_level""" , type=UpperCAmelCase__ , default="""O2""" , help=( """For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3'].""" """See details at https://nvidia.github.io/apex/amp.html""" ) , ) parser.add_argument("""--n_tpu_cores""" , dest="""tpu_cores""" , type=UpperCAmelCase__ ) parser.add_argument("""--max_grad_norm""" , dest="""gradient_clip_val""" , default=1.0 , type=UpperCAmelCase__ , help="""Max gradient norm""" ) parser.add_argument("""--do_train""" , action="""store_true""" , help="""Whether to run training.""" ) parser.add_argument("""--do_predict""" , action="""store_true""" , help="""Whether to run predictions on the test set.""" ) parser.add_argument( """--gradient_accumulation_steps""" , dest="""accumulate_grad_batches""" , type=UpperCAmelCase__ , default=1 , help="""Number of updates steps to accumulate before performing a backward/update pass.""" , ) parser.add_argument("""--seed""" , type=UpperCAmelCase__ , default=42 , help="""random seed for initialization""" ) parser.add_argument( """--data_dir""" , default=str(Path(UpperCAmelCase__ ).parent / """test_run""" / """dummy-train-data""" ) , type=UpperCAmelCase__ , help="""The input data dir. Should contain the training files for the CoNLL-2003 NER task.""" , ) def lowerCamelCase ( UpperCAmelCase__ : BaseTransformer , UpperCAmelCase__ : argparse.Namespace , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Tuple=[] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[Any]=None , **UpperCAmelCase__ : Tuple , ) -> Dict: pl.seed_everything(args.seed ) # init model lowercase_ : Optional[Any] = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=UpperCAmelCase__ ) # add custom checkpoints if checkpoint_callback is None: lowercase_ : int = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="""checkpoint""" , monitor="""val_loss""" , mode="""min""" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(UpperCAmelCase__ ) if logging_callback is None: lowercase_ : Optional[int] = LoggingCallback() lowercase_ : Any = {} if args.fpaa: lowercase_ : Union[str, Any] = 16 if args.gpus > 1: lowercase_ : Optional[Any] = """auto""" lowercase_ : Any = """ddp""" lowercase_ : List[Any] = args.accumulate_grad_batches lowercase_ : Union[str, Any] = None lowercase_ : List[str] = """auto""" lowercase_ : Any = pl.Trainer.from_argparse_args( UpperCAmelCase__ , weights_summary=UpperCAmelCase__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=UpperCAmelCase__ , val_check_interval=1 , num_sanity_val_steps=2 , **UpperCAmelCase__ , ) if args.do_train: trainer.fit(UpperCAmelCase__ ) else: print("""RAG modeling tests with new set functions successfuly executed!""" ) return trainer
239
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase : Dict = {"configuration_mmbt": ["MMBTConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict = ["MMBTForClassification", "MMBTModel", "ModalEmbeddings"] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys _lowercase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
239
1
"""simple docstring""" def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(_SCREAMING_SNAKE_CASE ) ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" if index == len(_SCREAMING_SNAKE_CASE ): return True # Recursive Step for i in range(_SCREAMING_SNAKE_CASE ): if valid_coloring(graph[index] , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): # Color current vertex UpperCamelCase = i # Validate coloring if util_color(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , index + 1 ): return True # Backtrack UpperCamelCase = -1 return False def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = [-1] * len(_SCREAMING_SNAKE_CASE ) if util_color(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 0 ): return colored_vertices return []
244
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowerCAmelCase__ = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' lowerCAmelCase__ = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' lowerCAmelCase__ = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return float((preds == labels).mean() ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="binary" ): """simple docstring""" UpperCamelCase = simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = float(fa_score(y_true=_SCREAMING_SNAKE_CASE , y_pred=_SCREAMING_SNAKE_CASE , average=_SCREAMING_SNAKE_CASE ) ) return { "accuracy": acc, "f1": fa, } def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = {} for id_pred, label in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase = F"{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}" UpperCamelCase = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCamelCase = [(pred, label)] UpperCamelCase , UpperCamelCase = [], [] for question, preds_labels in question_map.items(): UpperCamelCase , UpperCamelCase = zip(*_SCREAMING_SNAKE_CASE ) UpperCamelCase = fa_score(y_true=_SCREAMING_SNAKE_CASE , y_pred=_SCREAMING_SNAKE_CASE , average="macro" ) fas.append(_SCREAMING_SNAKE_CASE ) UpperCamelCase = int(sum(pred == label for pred, label in preds_labels ) == len(_SCREAMING_SNAKE_CASE ) ) ems.append(_SCREAMING_SNAKE_CASE ) UpperCamelCase = float(sum(_SCREAMING_SNAKE_CASE ) / len(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = sum(_SCREAMING_SNAKE_CASE ) / len(_SCREAMING_SNAKE_CASE ) UpperCamelCase = float(fa_score(y_true=_SCREAMING_SNAKE_CASE , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCamelCase ( datasets.Metric ): def snake_case_ (self ) -> Dict: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def snake_case_ (self ) -> Tuple: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def snake_case_ (self , __a , __a ) -> str: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__a , __a )} elif self.config_name == "cb": return acc_and_fa(__a , __a , fa_avg="macro" ) elif self.config_name == "record": UpperCamelCase = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] UpperCamelCase = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(__a , __a )[0] elif self.config_name == "multirc": return evaluate_multirc(__a , __a ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__a , __a )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
244
1
def UpperCamelCase_( lowerCamelCase_ ) -> list[int]: if length <= 0 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('Length must be a positive integer.' ) return [n * (2 * n - 1) for n in range(lowerCamelCase_ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
21
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ ={ 'configuration_roformer': ['ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoFormerConfig', 'RoFormerOnnxConfig'], 'tokenization_roformer': ['RoFormerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =['RoFormerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoFormerForCausalLM', 'RoFormerForMaskedLM', 'RoFormerForMultipleChoice', 'RoFormerForQuestionAnswering', 'RoFormerForSequenceClassification', 'RoFormerForTokenClassification', 'RoFormerLayer', 'RoFormerModel', 'RoFormerPreTrainedModel', 'load_tf_weights_in_roformer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRoFormerForCausalLM', 'TFRoFormerForMaskedLM', 'TFRoFormerForMultipleChoice', 'TFRoFormerForQuestionAnswering', 'TFRoFormerForSequenceClassification', 'TFRoFormerForTokenClassification', 'TFRoFormerLayer', 'TFRoFormerModel', 'TFRoFormerPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxRoFormerForMaskedLM', 'FlaxRoFormerForMultipleChoice', 'FlaxRoFormerForQuestionAnswering', 'FlaxRoFormerForSequenceClassification', 'FlaxRoFormerForTokenClassification', 'FlaxRoFormerModel', 'FlaxRoFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys lowercase__ =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
216
0
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A : List[Any] = logging.get_logger(__name__) A : Union[str, Any] = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } A : Optional[Any] = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __lowerCAmelCase ( a__ , a__ , a__ , a__ , a__ ) -> Dict: for attribute in key.split('''.''' ): __a = getattr(a__ , a__ ) if weight_type is not None: __a = getattr(a__ , a__ ).shape else: __a = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": __a = value elif weight_type == "weight_g": __a = value elif weight_type == "weight_v": __a = value elif weight_type == "bias": __a = value else: __a = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def __lowerCAmelCase ( a__ , a__ ) -> str: __a = [] __a = fairseq_model.state_dict() __a = hf_model.feature_extractor __a = hf_model.adapter for name, value in fairseq_dict.items(): __a = False if "conv_layers" in name: load_conv_layer( a__ , a__ , a__ , a__ , hf_model.config.feat_extract_norm == '''group''' , ) __a = True elif any(x in name for x in ['''adaptor''', '''w2v_encoder.proj.''', '''w2v_proj_ln.'''] ): load_adapter(a__ , a__ , a__ , a__ ) __a = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __a = True if "*" in mapped_key: __a = name.split(a__ )[0].split('''.''' )[-2] __a = mapped_key.replace('''*''' , a__ ) if "weight_g" in name: __a = '''weight_g''' elif "weight_v" in name: __a = '''weight_v''' elif "bias" in name: __a = '''bias''' elif "weight" in name: __a = '''weight''' else: __a = None set_recursively(a__ , a__ , a__ , a__ , a__ ) continue if not is_used: unused_weights.append(a__ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def __lowerCAmelCase ( a__ , a__ , a__ , a__ , a__ ) -> Any: __a = full_name.split('''conv_layers.''' )[-1] __a = name.split('''.''' ) __a = int(items[0] ) __a = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __a = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __a = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __a = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __a = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(a__ ) def __lowerCAmelCase ( a__ , a__ , a__ , a__ ) -> str: __a = full_name.split('''adaptor.''' )[-1] __a = name.split('''.''' ) if items[1].isdigit(): __a = int(items[1] ) else: __a = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.""" __a = value logger.info(F"""Adapter proj layer norm bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.""" __a = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.""" __a = value logger.info(F"""Adapter proj layer bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.""" __a = value logger.info(F"""Adapter proj layer weight was initialized from {full_name}.""" ) elif isinstance(a__ , a__ ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.""" __a = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.""" __a = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) else: unused_weights.append(a__ ) def __lowerCAmelCase ( a__ ) -> Optional[Any]: __a , __a = emb.weight.shape __a = nn.Linear(a__ , a__ , bias=a__ ) __a = emb.weight.data return lin_layer @torch.no_grad() def __lowerCAmelCase ( a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ) -> Union[str, Any]: __a = WavaVecaConfig.from_pretrained( a__ , add_adapter=a__ , adapter_stride=a__ , adapter_kernel_size=a__ , use_auth_token=a__ , output_hidden_size=a__ , ) __a = MBartConfig.from_pretrained(a__ ) # load model __a , __a , __a = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ '''config_yaml''': config_yaml_path, '''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path, '''load_pretrained_decoder_from''': None, } , ) __a = model[0].eval() # load feature extractor __a = WavaVecaFeatureExtractor.from_pretrained(a__ , use_auth_token=a__ ) # set weights for wav2vec2 encoder __a = WavaVecaModel(a__ ) recursively_load_weights_wavaveca(model.encoder , a__ ) # load decoder weights __a = MBartForCausalLM(a__ ) __a , __a = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=a__ ) logger.warning(F"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(F"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) __a = SpeechEncoderDecoderModel(encoder=a__ , decoder=a__ ) __a = False __a = MBartaaTokenizer(a__ ) tokenizer.save_pretrained(a__ ) __a = hf_wavavec.config.to_dict() __a = tokenizer.pad_token_id __a = tokenizer.bos_token_id __a = tokenizer.eos_token_id __a = '''mbart50''' __a = '''wav2vec2''' __a = tokenizer.eos_token_id __a = 25_0004 __a = tokenizer.eos_token_id __a = SpeechEncoderDecoderConfig.from_dict(a__ ) hf_wavavec.save_pretrained(a__ ) feature_extractor.save_pretrained(a__ ) if __name__ == "__main__": A : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_yaml_path', default=None, type=str, help='Path to yaml file of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-xls-r-1b', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/mbart-large-50-one-to-many-mmt', type=str, help='Path to hf decoder checkpoint config', ) parser.add_argument('--add_adapter', default=True, type=bool, help='whethere to add model adapter layers') parser.add_argument('--adapter_stride', default=2, type=int, help='stride of adapter layers') parser.add_argument('--adapter_kernel_size', default=3, type=int, help='kernel size of adapter layers') parser.add_argument('--encoder_output_dim', default=1_0_2_4, type=int, help='encoder output dim') parser.add_argument('--start_token_id', default=2_5_0_0_0_4, type=int, help='`decoder_start_token_id` of model config') A : Optional[int] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
33
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def __lowerCAmelCase ( a__ , a__ , a__ , a__ , a__ ) -> np.ndarray: __a = cva.getAffineTransform(a__ , a__ ) return cva.warpAffine(a__ , a__ , (rows, cols) ) if __name__ == "__main__": # read original image A : List[Any] = cva.imread( str(Path(__file__).resolve().parent.parent / 'image_data' / 'lena.jpg') ) # turn image in gray scale value A : Any = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape A , A : List[Any] = gray_img.shape # set different points to rotate image A : str = np.array([[5_0, 5_0], [2_0_0, 5_0], [5_0, 2_0_0]], np.floataa) A : Union[str, Any] = np.array([[1_0, 1_0_0], [2_0_0, 5_0], [1_0_0, 2_5_0]], np.floataa) A : Tuple = np.array([[5_0, 5_0], [1_5_0, 5_0], [1_2_0, 2_0_0]], np.floataa) A : Tuple = np.array([[1_0, 1_0_0], [8_0, 5_0], [1_8_0, 2_5_0]], np.floataa) # add all rotated images in a list A : Tuple = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations A : Union[str, Any] = plt.figure(1) A : str = ['Original', 'Rotation 1', 'Rotation 2', 'Rotation 3'] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, 'gray') plt.title(titles[i]) plt.axis('off') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
33
1