code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations from math import pi def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> dict[str, float]: '''simple docstring''' if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if inductance < 0: raise ValueError('''Inductance cannot be negative''' ) if frequency < 0: raise ValueError('''Frequency cannot be negative''' ) if reactance < 0: raise ValueError('''Inductive reactance cannot be negative''' ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = '▁' snake_case__ = {'vocab_file': 'vocab.txt', 'sentencepiece_model_ckpt': 'sentencepiece.bpe.model'} snake_case__ = { 'sentencepiece_model_file': 'sentencepiece.bpe.model', 'vocab_file': 'vocab.txt', } snake_case__ = { 'vocab_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', }, 'sentencepiece_model_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', }, } snake_case__ = { 'ernie-m-base': 514, 'ernie-m-large': 514, } snake_case__ = { 'ernie-m-base': {'do_lower_case': False}, 'ernie-m-large': {'do_lower_case': False}, } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ["input_ids"] A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_INIT_CONFIGURATION A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = RESOURCE_FILES_NAMES def __init__( self , A_ , A_=None , A_=False , A_="utf8" , A_="[UNK]" , A_="[SEP]" , A_="[PAD]" , A_="[CLS]" , A_="[MASK]" , A_ = None , **A_ , ) -> None: """simple docstring""" # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=A_ , unk_token=A_ , sep_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , vocab_file=A_ , encoding=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) _lowerCamelCase = do_lower_case _lowerCamelCase = sentencepiece_model_ckpt _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: _lowerCamelCase = self.load_vocab(filepath=A_ ) else: _lowerCamelCase = {self.sp_model.id_to_piece(A_ ): id for id in range(self.sp_model.get_piece_size() )} _lowerCamelCase = {v: k for k, v in self.vocab.items()} def UpperCamelCase_ ( self , A_ ) -> str: """simple docstring""" if text is None: return None _lowerCamelCase = self.tokenize(A_ ) _lowerCamelCase , _lowerCamelCase = '''''', [] for i, ch in enumerate(A_ ): if ch in self.SP_CHAR_MAPPING: _lowerCamelCase = self.SP_CHAR_MAPPING.get(A_ ) else: _lowerCamelCase = unicodedata.normalize('''NFKC''' , A_ ) if self.is_whitespace(A_ ): continue normalized_text += ch char_mapping.extend([i] * len(A_ ) ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = normalized_text, [], 0 if self.do_lower_case: _lowerCamelCase = text.lower() for token in split_tokens: if token[:1] == "▁": _lowerCamelCase = token[1:] _lowerCamelCase = text[offset:].index(A_ ) + offset _lowerCamelCase = start + len(A_ ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) _lowerCamelCase = end return token_mapping @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return len(self.vocab ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.__dict__.copy() _lowerCamelCase = None return state def __setstate__( self , A_ ) -> str: """simple docstring""" _lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _lowerCamelCase = {} _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" return "".join((self.SP_CHAR_MAPPING.get(A_ , A_ ) for c in text) ) def UpperCamelCase_ ( self , A_ , A_=False , A_=64 , A_=0.1 ) -> List[str]: """simple docstring""" if self.sp_model_kwargs.get('''enable_sampling''' ) is True: _lowerCamelCase = True if self.sp_model_kwargs.get('''alpha''' ) is not None: _lowerCamelCase = self.sp_model_kwargs.get('''alpha''' ) if self.sp_model_kwargs.get('''nbest_size''' ) is not None: _lowerCamelCase = self.sp_model_kwargs.get('''nbest_size''' ) if not enable_sampling: _lowerCamelCase = self.sp_model.EncodeAsPieces(A_ ) else: _lowerCamelCase = self.sp_model.SampleEncodeAsPieces(A_ , A_ , A_ ) _lowerCamelCase = [] for pi, piece in enumerate(A_ ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(A_ ) and pi != 0: new_pieces.append(A_ ) continue else: continue _lowerCamelCase = 0 for i, chunk in enumerate(A_ ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(A_ ) or self.is_punct(A_ ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(A_ ) _lowerCamelCase = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) _lowerCamelCase = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) _lowerCamelCase = i if len(A_ ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCamelCase_ ( self , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = ''''''.join(A_ ).replace(A_ , ''' ''' ).strip() return out_string def UpperCamelCase_ ( self , A_ ) -> Any: """simple docstring""" _lowerCamelCase = self.convert_ids_to_tokens(A_ ) _lowerCamelCase = ''''''.join(A_ ).replace(A_ , ''' ''' ).strip() return out_string def UpperCamelCase_ ( self , A_ ) -> List[str]: """simple docstring""" return self.vocab.get(A_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase_ ( self , A_ ) -> Union[str, Any]: """simple docstring""" return self.reverse_vocab.get(A_ , self.unk_token ) def UpperCamelCase_ ( self , A_ , A_=None ) -> List[str]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] _lowerCamelCase = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCamelCase_ ( self , A_ , A_=None ) -> List[Any]: """simple docstring""" if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCamelCase_ ( self , A_ , A_=None , A_=False ) -> List[str]: """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(A_ )) + [1, 1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1] def UpperCamelCase_ ( self , A_ , A_ = None ) -> List[int]: """simple docstring""" # called when `add_special_tokens` is True, so align with `build_inputs_with_special_tokens` method if token_ids_a is None: # [CLS] X [SEP] return (len(A_ ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(A_ ) + 1) + [1] * (len(A_ ) + 3) def UpperCamelCase_ ( self , A_ ) -> Dict: """simple docstring""" if "\u4e00" <= char <= "\u9fff": return True return False def UpperCamelCase_ ( self , A_ ) -> Optional[Any]: """simple docstring""" if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCamelCase_ ( self , A_ ) -> Union[str, Any]: """simple docstring""" if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(A_ ) == 1: _lowerCamelCase = unicodedata.category(A_ ) if cat == "Zs": return True return False def UpperCamelCase_ ( self , A_ ) -> Tuple: """simple docstring""" _lowerCamelCase = {} with io.open(A_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(A_ ): _lowerCamelCase = line.rstrip('''\n''' ) _lowerCamelCase = int(A_ ) return token_to_idx def UpperCamelCase_ ( self , A_ , A_ = None ) -> Tuple[str]: """simple docstring""" _lowerCamelCase = 0 if os.path.isdir(A_ ): _lowerCamelCase = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: _lowerCamelCase = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory with open(A_ , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda A_ : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ''' Please check that the vocabulary is not corrupted!''' ) _lowerCamelCase = token_index writer.write(token + '''\n''' ) index += 1 _lowerCamelCase = os.path.join(A_ , '''sentencepiece.bpe.model''' ) with open(A_ , '''wb''' ) as fi: _lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(A_ ) return (vocab_file,)
638
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
1
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
1
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> Any: '''simple docstring''' _lowerCamelCase = OmegaConf.load(__UpperCAmelCase ) if display: print(yaml.dump(OmegaConf.to_container(__UpperCAmelCase ) ) ) return config def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None ) -> List[Any]: '''simple docstring''' if conf_path is None: _lowerCamelCase = '''./model_checkpoints/vqgan_only.yaml''' _lowerCamelCase = load_config(__UpperCAmelCase , display=__UpperCAmelCase ) _lowerCamelCase = VQModel(**config.model.params ) if ckpt_path is None: _lowerCamelCase = '''./model_checkpoints/vqgan_only.pt''' _lowerCamelCase = torch.load(__UpperCAmelCase , map_location=__UpperCAmelCase ) if ".ckpt" in ckpt_path: _lowerCamelCase = sd['''state_dict'''] model.load_state_dict(__UpperCAmelCase , strict=__UpperCAmelCase ) model.to(__UpperCAmelCase ) del sd return model def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = model.encode(__UpperCAmelCase ) print(F'VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}' ) _lowerCamelCase = model.decode(__UpperCAmelCase ) return xrec def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[str]: '''simple docstring''' _lowerCamelCase , _lowerCamelCase = string.rsplit('''.''' , 1 ) if reload: _lowerCamelCase = importlib.import_module(__UpperCAmelCase ) importlib.reload(__UpperCAmelCase ) return getattr(importlib.import_module(__UpperCAmelCase , package=__UpperCAmelCase ) , cls ) def __magic_name__( __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' if "target" not in config: raise KeyError('''Expected key `target` to instantiate.''' ) return get_obj_from_str(config['''target'''] )(**config.get('''params''' , {} ) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True ) -> List[str]: '''simple docstring''' _lowerCamelCase = instantiate_from_config(__UpperCAmelCase ) if sd is not None: model.load_state_dict(__UpperCAmelCase ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' if ckpt: _lowerCamelCase = torch.load(__UpperCAmelCase , map_location='''cpu''' ) _lowerCamelCase = pl_sd['''global_step'''] print(F'loaded model from global step {global_step}.' ) else: _lowerCamelCase = {'''state_dict''': None} _lowerCamelCase = None _lowerCamelCase = load_model_from_config(config.model , pl_sd['''state_dict'''] , gpu=__UpperCAmelCase , eval_mode=__UpperCAmelCase )['''model'''] return model, global_step
638
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def __magic_name__( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase = SwinvaConfig() _lowerCamelCase = swinva_name.split('''_''' ) _lowerCamelCase = name_split[1] if "to" in name_split[3]: _lowerCamelCase = int(name_split[3][-3:] ) else: _lowerCamelCase = int(name_split[3] ) if "to" in name_split[2]: _lowerCamelCase = int(name_split[2][-2:] ) else: _lowerCamelCase = int(name_split[2][6:] ) if model_size == "tiny": _lowerCamelCase = 96 _lowerCamelCase = (2, 2, 6, 2) _lowerCamelCase = (3, 6, 12, 24) elif model_size == "small": _lowerCamelCase = 96 _lowerCamelCase = (2, 2, 18, 2) _lowerCamelCase = (3, 6, 12, 24) elif model_size == "base": _lowerCamelCase = 128 _lowerCamelCase = (2, 2, 18, 2) _lowerCamelCase = (4, 8, 16, 32) else: _lowerCamelCase = 192 _lowerCamelCase = (2, 2, 18, 2) _lowerCamelCase = (6, 12, 24, 48) if "to" in swinva_name: _lowerCamelCase = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): _lowerCamelCase = 2_1841 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-22k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} else: _lowerCamelCase = 1000 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = img_size _lowerCamelCase = num_classes _lowerCamelCase = embed_dim _lowerCamelCase = depths _lowerCamelCase = num_heads _lowerCamelCase = window_size return config def __magic_name__( __UpperCAmelCase ) -> List[str]: '''simple docstring''' if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if "q_bias" in name: _lowerCamelCase = name.replace('''q_bias''' , '''query.bias''' ) if "k_bias" in name: _lowerCamelCase = name.replace('''k_bias''' , '''key.bias''' ) if "v_bias" in name: _lowerCamelCase = name.replace('''v_bias''' , '''value.bias''' ) if "cpb_mlp" in name: _lowerCamelCase = name.replace('''cpb_mlp''' , '''continuous_position_bias_mlp''' ) if name == "norm.weight": _lowerCamelCase = '''layernorm.weight''' if name == "norm.bias": _lowerCamelCase = '''layernorm.bias''' if "head" in name: _lowerCamelCase = name.replace('''head''' , '''classifier''' ) else: _lowerCamelCase = '''swinv2.''' + name return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "mask" in key: continue elif "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[1] ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[ dim : dim * 2 ] _lowerCamelCase = val[-dim:] else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ) timm_model.eval() _lowerCamelCase = get_swinva_config(__UpperCAmelCase ) _lowerCamelCase = SwinvaForImageClassification(__UpperCAmelCase ) model.eval() _lowerCamelCase = convert_state_dict(timm_model.state_dict() , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = AutoImageProcessor.from_pretrained('''microsoft/{}'''.format(swinva_name.replace('''_''' , '''-''' ) ) ) _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) _lowerCamelCase = image_processor(images=__UpperCAmelCase , return_tensors='''pt''' ) _lowerCamelCase = timm_model(inputs['''pixel_values'''] ) _lowerCamelCase = model(**__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print(F'Saving model {swinva_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) model.push_to_hub( repo_path_or_name=Path(__UpperCAmelCase , __UpperCAmelCase ) , organization='''nandwalritik''' , commit_message='''Add model''' , ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swinv2_name', default='swinv2_tiny_patch4_window8_256', type=str, help='Name of the Swinv2 timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) snake_case__ = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
638
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } _lowerCamelCase = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub _lowerCamelCase = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase_ ( self , **A_ ) -> str: """simple docstring""" _lowerCamelCase = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> Optional[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> int: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = feature_extractor(A_ , return_tensors='''np''' ) _lowerCamelCase = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = '''This is a test string''' _lowerCamelCase = processor(text=A_ ) _lowerCamelCase = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase_ ( self , A_=(2, 10, 16) , A_=77 ) -> Optional[Any]: """simple docstring""" np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits(shape=(10, 16) , seed=13 ) _lowerCamelCase = processor.decode(A_ ) _lowerCamelCase = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: _lowerCamelCase = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: _lowerCamelCase = processor.batch_decode(A_ , A_ ) _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as p: _lowerCamelCase = decoder.decode_beams_batch(A_ , A_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 15 _lowerCamelCase = -20.0 _lowerCamelCase = -4.0 _lowerCamelCase = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] _lowerCamelCase = [d[0][2] for d in decoded_decoder_out] _lowerCamelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 2.0 _lowerCamelCase = 5.0 _lowerCamelCase = -20.0 _lowerCamelCase = True _lowerCamelCase = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = snapshot_download('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = processor_wavaveca(A_ , return_tensors='''np''' ) _lowerCamelCase = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor_wavaveca.batch_decode(A_ ) _lowerCamelCase = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = [d[key] for d in offsets] return retrieved_list def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits()[0] _lowerCamelCase = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" import torch _lowerCamelCase = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) _lowerCamelCase = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) _lowerCamelCase = iter(A_ ) _lowerCamelCase = next(A_ ) _lowerCamelCase = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) _lowerCamelCase = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train _lowerCamelCase = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): _lowerCamelCase = model(A_ ).logits.cpu().numpy() _lowerCamelCase = processor.decode(logits[0] , output_word_offsets=A_ ) _lowerCamelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate _lowerCamelCase = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] _lowerCamelCase = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off _lowerCamelCase = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) _lowerCamelCase = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
638
1
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , A_ = None , A_ = None , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , **A_ , ) -> str: """simple docstring""" _lowerCamelCase = path_or_paths _lowerCamelCase = split if split or isinstance(A_ , A_ ) else '''train''' _lowerCamelCase = features _lowerCamelCase = cache_dir _lowerCamelCase = keep_in_memory _lowerCamelCase = streaming _lowerCamelCase = num_proc _lowerCamelCase = kwargs @abstractmethod def UpperCamelCase_ ( self ) -> Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]: """simple docstring""" pass class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , **A_ , ) -> Any: """simple docstring""" _lowerCamelCase = features _lowerCamelCase = cache_dir _lowerCamelCase = keep_in_memory _lowerCamelCase = streaming _lowerCamelCase = num_proc _lowerCamelCase = kwargs @abstractmethod def UpperCamelCase_ ( self ) -> Union[Dataset, IterableDataset]: """simple docstring""" pass
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _lowerCamelCase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _lowerCamelCase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _lowerCamelCase = subset[i - 1][j] if arr[i - 1] <= j: _lowerCamelCase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
638
1
import requests from bsa import BeautifulSoup def __magic_name__( __UpperCAmelCase = "https://www.worldometers.info/coronavirus" ) -> dict: '''simple docstring''' _lowerCamelCase = BeautifulSoup(requests.get(__UpperCAmelCase ).text , '''html.parser''' ) _lowerCamelCase = soup.findAll('''h1''' ) _lowerCamelCase = soup.findAll('''div''' , {'''class''': '''maincounter-number'''} ) keys += soup.findAll('''span''' , {'''class''': '''panel-title'''} ) values += soup.findAll('''div''' , {'''class''': '''number-table-main'''} ) return {key.text.strip(): value.text.strip() for key, value in zip(__UpperCAmelCase , __UpperCAmelCase )} if __name__ == "__main__": print('\033[1m' + 'COVID-19 Status of the World' + '\033[0m\n') for key, value in world_covidaa_stats().items(): print(f'''{key}\n{value}\n''')
638
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class UpperCamelCase ( __lowercase ): '''simple docstring''' def UpperCamelCase_ ( self , A_ ) -> Union[str, Any]: """simple docstring""" with open(A_ , encoding='''utf-8''' ) as input_file: _lowerCamelCase = re.compile(r'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''' ) _lowerCamelCase = input_file.read() _lowerCamelCase = regexp.search(A_ ) return match def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" with open(A_ , encoding='''utf-8''' ) as input_file: _lowerCamelCase = re.compile(r'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''' , re.DOTALL ) _lowerCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _lowerCamelCase = regexp.finditer(A_ ) _lowerCamelCase = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = Path('''./datasets''' ) _lowerCamelCase = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(A_ ) ): raise AssertionError(F'open(...) must use utf-8 encoding in {dataset}' ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = Path('''./datasets''' ) _lowerCamelCase = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_print_statements(str(A_ ) ): raise AssertionError(F'print statement found in {dataset}. Use datasets.logger/logging instead.' )
638
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
1
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = (DDPMParallelScheduler,) def UpperCamelCase_ ( self , **A_ ) -> Tuple: """simple docstring""" _lowerCamelCase = { '''num_train_timesteps''': 10_00, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''variance_type''': '''fixed_small''', '''clip_sample''': True, } config.update(**A_ ) return config def UpperCamelCase_ ( self ) -> int: """simple docstring""" for timesteps in [1, 5, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=A_ , beta_end=A_ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=A_ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" for clip_sample in [True, False]: self.check_over_configs(clip_sample=A_ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" self.check_over_configs(thresholding=A_ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=A_ , prediction_type=A_ , sample_max_value=A_ , ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" for t in [0, 5_00, 9_99]: self.check_over_forward(time_step=A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 ) - 0.00979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 ) - 0.02 ) ) < 1E-5 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = len(A_ ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter _lowerCamelCase = self.dummy_sample_deter + 0.1 _lowerCamelCase = self.dummy_sample_deter - 0.1 _lowerCamelCase = samplea.shape[0] _lowerCamelCase = torch.stack([samplea, samplea, samplea] , dim=0 ) _lowerCamelCase = torch.arange(A_ )[0:3, None].repeat(1 , A_ ) _lowerCamelCase = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) _lowerCamelCase = scheduler.batch_step_no_noise(A_ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 1153.1833 ) < 1E-2 assert abs(result_mean.item() - 0.5005 ) < 1E-3 def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = len(A_ ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter _lowerCamelCase = torch.manual_seed(0 ) for t in reversed(range(A_ ) ): # 1. predict noise residual _lowerCamelCase = model(A_ , A_ ) # 2. predict previous mean of sample x_t-1 _lowerCamelCase = scheduler.step(A_ , A_ , A_ , generator=A_ ).prev_sample _lowerCamelCase = pred_prev_sample _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config(prediction_type='''v_prediction''' ) _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = len(A_ ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter _lowerCamelCase = torch.manual_seed(0 ) for t in reversed(range(A_ ) ): # 1. predict noise residual _lowerCamelCase = model(A_ , A_ ) # 2. predict previous mean of sample x_t-1 _lowerCamelCase = scheduler.step(A_ , A_ , A_ , generator=A_ ).prev_sample _lowerCamelCase = pred_prev_sample _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = [1_00, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=A_ ) _lowerCamelCase = scheduler.timesteps for i, timestep in enumerate(A_ ): if i == len(A_ ) - 1: _lowerCamelCase = -1 else: _lowerCamelCase = timesteps[i + 1] _lowerCamelCase = scheduler.previous_timestep(A_ ) _lowerCamelCase = prev_t.item() self.assertEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = [1_00, 87, 50, 51, 0] with self.assertRaises(A_ , msg='''`custom_timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=A_ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = [1_00, 87, 50, 1, 0] _lowerCamelCase = len(A_ ) with self.assertRaises(A_ , msg='''Can only pass one of `num_inference_steps` or `custom_timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=A_ , timesteps=A_ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) _lowerCamelCase = [scheduler.config.num_train_timesteps] with self.assertRaises( A_ , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=A_ )
638
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
1
import numpy as np def __magic_name__( __UpperCAmelCase ) -> np.array: '''simple docstring''' return 1 / (1 + np.exp(-vector )) if __name__ == "__main__": import doctest doctest.testmod()
638
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ) -> None: """simple docstring""" _lowerCamelCase = device _lowerCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _lowerCamelCase = [0.48145466, 0.4578275, 0.40821073] _lowerCamelCase = [0.26862954, 0.26130258, 0.27577711] _lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCamelCase = torchvision.transforms.Resize(2_24 ) _lowerCamelCase = torchvision.transforms.CenterCrop(2_24 ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.resize(A_ ) _lowerCamelCase = self.center_crop(A_ ) _lowerCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer(text=A_ , **A_ ) _lowerCamelCase = self.preprocess_img(A_ ) _lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ) -> None: """simple docstring""" super().__init__() _lowerCamelCase = None _lowerCamelCase = device if device else get_device() if vqgan: _lowerCamelCase = vqgan else: _lowerCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _lowerCamelCase = clip else: _lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _lowerCamelCase = ProcessorGradientFlow(device=self.device ) _lowerCamelCase = iterations _lowerCamelCase = lr _lowerCamelCase = log _lowerCamelCase = make_grid _lowerCamelCase = return_val _lowerCamelCase = quantize _lowerCamelCase = self.vqgan.decoder.z_shape def UpperCamelCase_ ( self , A_=None , A_=None , A_=5 , A_=True ) -> Any: """simple docstring""" _lowerCamelCase = [] if output_path is None: _lowerCamelCase = '''./animation.gif''' if input_path is None: _lowerCamelCase = self.save_path _lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(A_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(A_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _lowerCamelCase = total_duration / len(A_ ) _lowerCamelCase = [frame_duration] * len(A_ ) if extend_frames: _lowerCamelCase = 1.5 _lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F'gif saved to {output_path}' ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> Union[str, Any]: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _lowerCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _lowerCamelCase = preprocess_vqgan(A_ ) _lowerCamelCase , *_lowerCamelCase = self.vqgan.encode(A_ ) return z def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.latent.detach().requires_grad_() _lowerCamelCase = base_latent + transform_vector if self.quantize: _lowerCamelCase , *_lowerCamelCase = self.vqgan.quantize(A_ ) else: _lowerCamelCase = trans_latent return self.vqgan.decode(A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_=None ) -> Any: """simple docstring""" _lowerCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors='''pt''' , padding=A_ ) _lowerCamelCase = self.clip(**A_ ) _lowerCamelCase = clip_outputs.logits_per_image if weights is not None: _lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , A_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , A_ , weights=neg_prompts['''weights'''] ) else: _lowerCamelCase = torch.tensor([1] , device=self.device ) _lowerCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCamelCase = self._add_vector(A_ ) _lowerCamelCase = loop_post_process(A_ ) _lowerCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print('''CLIP loss''' , A_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" wandb.init(reinit=A_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _lowerCamelCase = Image.open(A_ ) _lowerCamelCase = image.resize((2_56, 2_56) ) wandb.log('''Original Image''' , wandb.Image(A_ ) ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if not prompts: return [] _lowerCamelCase = [] _lowerCamelCase = [] if isinstance(A_ , A_ ): _lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _lowerCamelCase = prompt[0] _lowerCamelCase = float(prompt[1] ) elif ":" in prompt: _lowerCamelCase , _lowerCamelCase = prompt.split(''':''' ) _lowerCamelCase = float(A_ ) else: _lowerCamelCase = prompt _lowerCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ) -> str: """simple docstring""" if image_path: _lowerCamelCase = self._get_latent(A_ ) else: _lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCamelCase = self.process_prompts(A_ ) _lowerCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(A_ ) _lowerCamelCase = save_path _lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(A_ ) ) _lowerCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({'''Image''': wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
638
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) snake_case__ = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case__ = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "" , A_ = False ) -> None: """simple docstring""" # Mapping from the first character of the prefix of the node _lowerCamelCase = {} # A node will be a leaf if the tree contains its word _lowerCamelCase = is_leaf _lowerCamelCase = prefix def UpperCamelCase_ ( self , A_ ) -> tuple[str, str, str]: """simple docstring""" _lowerCamelCase = 0 for q, w in zip(self.prefix , A_ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" for word in words: self.insert(A_ ) def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" # Case 1: If the word is the prefix of the node # Solution: We set the current node as leaf if self.prefix == word: _lowerCamelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _lowerCamelCase = RadixNode(prefix=A_ , is_leaf=A_ ) else: _lowerCamelCase = self.nodes[word[0]] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = incoming_node.match( A_ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(A_ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _lowerCamelCase = remaining_prefix _lowerCamelCase = self.nodes[matching_string[0]] _lowerCamelCase = RadixNode(A_ , A_ ) _lowerCamelCase = aux_node if remaining_word == "": _lowerCamelCase = True else: self.nodes[matching_string[0]].insert(A_ ) def UpperCamelCase_ ( self , A_ ) -> bool: """simple docstring""" _lowerCamelCase = self.nodes.get(word[0] , A_ ) if not incoming_node: return False else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = incoming_node.match( A_ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(A_ ) def UpperCamelCase_ ( self , A_ ) -> bool: """simple docstring""" _lowerCamelCase = self.nodes.get(word[0] , A_ ) if not incoming_node: return False else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = incoming_node.match( A_ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(A_ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _lowerCamelCase = list(self.nodes.values() )[0] _lowerCamelCase = merging_node.is_leaf self.prefix += merging_node.prefix _lowerCamelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _lowerCamelCase = False # If there is 1 edge, we merge it with its child else: _lowerCamelCase = list(incoming_node.nodes.values() )[0] _lowerCamelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _lowerCamelCase = merging_node.nodes return True def UpperCamelCase_ ( self , A_ = 0 ) -> None: """simple docstring""" if self.prefix != "": print('''-''' * height , self.prefix , ''' (leaf)''' if self.is_leaf else '''''' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __magic_name__( ) -> bool: '''simple docstring''' _lowerCamelCase = '''banana bananas bandana band apple all beast'''.split() _lowerCamelCase = RadixNode() root.insert_many(__UpperCAmelCase ) assert all(root.find(__UpperCAmelCase ) for word in words ) assert not root.find('''bandanas''' ) assert not root.find('''apps''' ) root.delete('''all''' ) assert not root.find('''all''' ) root.delete('''banana''' ) assert not root.find('''banana''' ) assert root.find('''bananas''' ) return True def __magic_name__( ) -> None: '''simple docstring''' assert test_trie() def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = RadixNode() _lowerCamelCase = '''banana bananas bandanas bandana band apple all beast'''.split() root.insert_many(__UpperCAmelCase ) print('''Words:''' , __UpperCAmelCase ) print('''Tree:''' ) root.print_tree() if __name__ == "__main__": main()
638
import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='%(message)s') def __magic_name__( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size, 1) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.mean(1 ) # Centralize the data of class i _lowerCamelCase = data - column_reshape(__UpperCAmelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(__UpperCAmelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = features.mean(1 ) _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.shape[1] _lowerCamelCase = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' if features.any(): _lowerCamelCase = features.mean(1 ) # Center the dataset _lowerCamelCase = features - np.reshape(__UpperCAmelCase , (data_mean.size, 1) ) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) / features.shape[1] _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(__UpperCAmelCase ) # Take all the columns in the reverse order (-1), and then takes only the first _lowerCamelCase = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space _lowerCamelCase = np.dot(filtered_eigenvectors.T , __UpperCAmelCase ) logging.info('''Principal Component Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' assert classes > dimensions # Check if features have been already loaded if features.any: _lowerCamelCase , _lowerCamelCase = eigh( covariance_between_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , covariance_within_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , ) _lowerCamelCase = eigenvectors[:, ::-1][:, :dimensions] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = np.linalg.svd(__UpperCAmelCase ) _lowerCamelCase = svd_matrix[:, 0:dimensions] _lowerCamelCase = np.dot(filtered_svd_matrix.T , __UpperCAmelCase ) logging.info('''Linear Discriminant Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) _lowerCamelCase = np.array([0, 0, 0, 1, 1] ) _lowerCamelCase = 2 _lowerCamelCase = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = linear_discriminant_analysis( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if isinstance(__UpperCAmelCase , np.ndarray ): raise AssertionError( '''Did not raise AssertionError for dimensions > classes''' ) assert error_info.type is AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) _lowerCamelCase = 2 _lowerCamelCase = np.array([[6.9_2_8_2_0_3_2_3, 8.6_6_0_2_5_4_0_4, 1_0.3_9_2_3_0_4_8_5], [3.0, 3.0, 3.0]] ) with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = principal_component_analysis(__UpperCAmelCase , __UpperCAmelCase ) if not np.allclose(__UpperCAmelCase , __UpperCAmelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
638
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = StableDiffusionXLImgaImgPipeline A_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} A_ = PipelineTesterMixin.required_optional_params - {'latents'} A_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ = IMAGE_TO_IMAGE_IMAGE_PARAMS A_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=A_ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=32 , ) _lowerCamelCase = CLIPTextModel(A_ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=A_ ) _lowerCamelCase = CLIPTextModelWithProjection(A_ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=A_ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def UpperCamelCase_ ( self , A_ , A_=0 ) -> List[str]: """simple docstring""" _lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_ ) ).to(A_ ) _lowerCamelCase = image / 2 + 0.5 if str(A_ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(A_ ) else: _lowerCamelCase = torch.Generator(device=A_ ).manual_seed(A_ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.75, } return inputs def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**A_ ) _lowerCamelCase = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = self.get_dummy_inputs(A_ ) _lowerCamelCase = sd_pipe(**A_ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**A_ ) _lowerCamelCase = sd_pipe.to(A_ ) _lowerCamelCase = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(A_ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**A_ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(A_ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(A_ , negative_prompt=A_ ) _lowerCamelCase = sd_pipe( **A_ , prompt_embeds=A_ , negative_prompt_embeds=A_ , pooled_prompt_embeds=A_ , negative_pooled_prompt_embeds=A_ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self , A_ , A_="cpu" , A_=torch.floataa , A_=0 ) -> Tuple: """simple docstring""" _lowerCamelCase = torch.Generator(device=A_ ).manual_seed(A_ ) _lowerCamelCase = np.random.RandomState(A_ ).standard_normal((1, 4, 64, 64) ) _lowerCamelCase = torch.from_numpy(A_ ).to(device=A_ , dtype=A_ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = self.get_inputs(A_ ) _lowerCamelCase = pipe(**A_ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) _lowerCamelCase = np.array([0.49493, 0.47896, 0.40798, 0.54214, 0.53212, 0.48202, 0.47656, 0.46329, 0.48506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
638
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['vqvae'] def __init__( self , A_ , A_ , A_ , A_ , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules(unet=A_ , scheduler=A_ , mel=A_ , vqvae=A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 50 if isinstance(self.scheduler , A_ ) else 10_00 @torch.no_grad() def __call__( self , A_ = 1 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = 0 , A_ = None , A_ = None , A_=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: """simple docstring""" _lowerCamelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(A_ ) _lowerCamelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: _lowerCamelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: _lowerCamelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=A_ , device=self.device , ) _lowerCamelCase = noise _lowerCamelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A_ , A_ ) _lowerCamelCase = self.mel.audio_slice_to_image(A_ ) _lowerCamelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) _lowerCamelCase = (input_image / 2_55) * 2 - 1 _lowerCamelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: _lowerCamelCase = self.vqvae.encode(torch.unsqueeze(A_ , 0 ) ).latent_dist.sample( generator=A_ )[0] _lowerCamelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , self.scheduler.timesteps[start_step - 1] ) _lowerCamelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) _lowerCamelCase = int(mask_start_secs * pixels_per_second ) _lowerCamelCase = int(mask_end_secs * pixels_per_second ) _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , A_ ): _lowerCamelCase = self.unet(A_ , A_ , A_ )['''sample'''] else: _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] if isinstance(self.scheduler , A_ ): _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , eta=A_ , generator=A_ , )['''prev_sample'''] else: _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , generator=A_ , )['''prev_sample'''] if mask is not None: if mask_start > 0: _lowerCamelCase = mask[:, step, :, :mask_start] if mask_end > 0: _lowerCamelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance _lowerCamelCase = 1 / self.vqvae.config.scaling_factor * images _lowerCamelCase = self.vqvae.decode(A_ )['''sample'''] _lowerCamelCase = (images / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() _lowerCamelCase = (images * 2_55).round().astype('''uint8''' ) _lowerCamelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A_ , mode='''RGB''' ).convert('''L''' ) for _ in images) ) _lowerCamelCase = [self.mel.image_to_audio(A_ ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A_ )[:, np.newaxis, :] ) , **ImagePipelineOutput(A_ ) ) @torch.no_grad() def UpperCamelCase_ ( self , A_ , A_ = 50 ) -> np.ndarray: """simple docstring""" assert isinstance(self.scheduler , A_ ) self.scheduler.set_timesteps(A_ ) _lowerCamelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) _lowerCamelCase = (sample / 2_55) * 2 - 1 _lowerCamelCase = torch.Tensor(A_ ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): _lowerCamelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps _lowerCamelCase = self.scheduler.alphas_cumprod[t] _lowerCamelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) _lowerCamelCase = 1 - alpha_prod_t _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] _lowerCamelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output _lowerCamelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) _lowerCamelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def UpperCamelCase_ ( A_ , A_ , A_ ) -> torch.Tensor: """simple docstring""" _lowerCamelCase = acos(torch.dot(torch.flatten(A_ ) , torch.flatten(A_ ) ) / torch.norm(A_ ) / torch.norm(A_ ) ) return sin((1 - alpha) * theta ) * xa / sin(A_ ) + sin(alpha * theta ) * xa / sin(A_ )
638
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __magic_name__( __UpperCAmelCase ) -> Dict: '''simple docstring''' return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __magic_name__( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = create_tensor(__UpperCAmelCase ) _lowerCamelCase = gather(__UpperCAmelCase ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def __magic_name__( __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [state.process_index] _lowerCamelCase = gather_object(__UpperCAmelCase ) assert len(__UpperCAmelCase ) == state.num_processes, F'{gathered_obj}, {len(__UpperCAmelCase )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __magic_name__( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = create_tensor(__UpperCAmelCase ) _lowerCamelCase = broadcast(__UpperCAmelCase ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def __magic_name__( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if state.is_main_process: _lowerCamelCase = torch.arange(state.num_processes + 1 ).to(state.device ) else: _lowerCamelCase = torch.arange(state.num_processes ).to(state.device ) _lowerCamelCase = pad_across_processes(__UpperCAmelCase ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def __magic_name__( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if state.num_processes != 2: return _lowerCamelCase = create_tensor(__UpperCAmelCase ) _lowerCamelCase = reduce(__UpperCAmelCase , '''sum''' ) _lowerCamelCase = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase ), F'{reduced_tensor} != {truth_tensor}' def __magic_name__( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if state.num_processes != 2: return _lowerCamelCase = create_tensor(__UpperCAmelCase ) _lowerCamelCase = reduce(__UpperCAmelCase , '''mean''' ) _lowerCamelCase = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase ), F'{reduced_tensor} != {truth_tensor}' def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' main() def __magic_name__( ) -> int: '''simple docstring''' _lowerCamelCase = PartialState() state.print(F'State: {state}' ) state.print('''testing gather''' ) test_gather(__UpperCAmelCase ) state.print('''testing gather_object''' ) test_gather_object(__UpperCAmelCase ) state.print('''testing broadcast''' ) test_broadcast(__UpperCAmelCase ) state.print('''testing pad_across_processes''' ) test_pad_across_processes(__UpperCAmelCase ) state.print('''testing reduce_sum''' ) test_reduce_sum(__UpperCAmelCase ) state.print('''testing reduce_mean''' ) test_reduce_mean(__UpperCAmelCase ) if __name__ == "__main__": main()
638
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowercase ) , 'Tatoeba directory does not exist.' ) class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=A_ ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=A_ ) assert mmeta["long_pair"] == "heb-eng"
638
1
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration snake_case__ = HfArgumentParser(InitializationArguments) snake_case__ = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization snake_case__ = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks snake_case__ = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) snake_case__ = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config snake_case__ = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
638
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' _lowerCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase = in_proj_bias[: config.hidden_size] _lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase = in_proj_bias[-config.hidden_size :] def __magic_name__( __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True ) -> str: '''simple docstring''' _lowerCamelCase = ViTConfig() # patch_size if model_name[-1] == "8": _lowerCamelCase = 8 # set labels if required if not base_model: _lowerCamelCase = 1000 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _lowerCamelCase = 384 _lowerCamelCase = 1536 _lowerCamelCase = 12 _lowerCamelCase = 6 # load original model from torch hub _lowerCamelCase = torch.hub.load('''facebookresearch/dino:main''' , __UpperCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase = original_model.state_dict() if base_model: remove_classification_head_(__UpperCAmelCase ) _lowerCamelCase = create_rename_keys(__UpperCAmelCase , base_model=__UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # load HuggingFace model if base_model: _lowerCamelCase = ViTModel(__UpperCAmelCase , add_pooling_layer=__UpperCAmelCase ).eval() else: _lowerCamelCase = ViTForImageClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _lowerCamelCase = ViTImageProcessor() _lowerCamelCase = image_processor(images=prepare_img() , return_tensors='''pt''' ) _lowerCamelCase = encoding['''pixel_values'''] _lowerCamelCase = model(__UpperCAmelCase ) if base_model: _lowerCamelCase = original_model(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _lowerCamelCase = original_model(__UpperCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) snake_case__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
638
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'yolos' def __init__( self , A_=7_68 , A_=12 , A_=12 , A_=30_72 , A_="gelu" , A_=0.0 , A_=0.0 , A_=0.02 , A_=1E-1_2 , A_=[5_12, 8_64] , A_=16 , A_=3 , A_=True , A_=1_00 , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=5 , A_=2 , A_=0.1 , **A_ , ) -> str: """simple docstring""" super().__init__(**A_ ) _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = qkv_bias _lowerCamelCase = num_detection_tokens _lowerCamelCase = use_mid_position_embeddings _lowerCamelCase = auxiliary_loss # Hungarian matcher _lowerCamelCase = class_cost _lowerCamelCase = bbox_cost _lowerCamelCase = giou_cost # Loss coefficients _lowerCamelCase = bbox_loss_coefficient _lowerCamelCase = giou_loss_coefficient _lowerCamelCase = eos_coefficient class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = version.parse('1.11' ) @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def UpperCamelCase_ ( self ) -> float: """simple docstring""" return 1E-4 @property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 12
638
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _lowerCamelCase = '''lm_head''' _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): _lowerCamelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True ) -> Union[str, Any]: '''simple docstring''' if config_path is not None: _lowerCamelCase = UniSpeechConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _lowerCamelCase = Dictionary.load_from_json(__UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _lowerCamelCase = target_dict.pad_index _lowerCamelCase = target_dict.bos_index _lowerCamelCase = target_dict.eos_index _lowerCamelCase = len(target_dict.symbols ) _lowerCamelCase = os.path.join(__UpperCAmelCase , '''vocab.json''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__UpperCAmelCase ) ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) _lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _lowerCamelCase = 42 _lowerCamelCase = 43 with open(__UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = WavaVecaPhonemeCTCTokenizer( __UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__UpperCAmelCase , ) _lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False _lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) _lowerCamelCase = WavaVecaProcessor(feature_extractor=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) _lowerCamelCase = UniSpeechForCTC(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechForPreTraining(__UpperCAmelCase ) if is_finetuned: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _lowerCamelCase = model[0].eval() recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) hf_unispeech.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
638
1
import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin snake_case__ = logging.get_logger(__name__) enable_full_determinism() class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = UNetaDModel A_ = 'sample' @property def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = 4 _lowerCamelCase = 3 _lowerCamelCase = (32, 32) _lowerCamelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) _lowerCamelCase = torch.tensor([10] ).to(A_ ) return {"sample": noise, "timestep": time_step} @property def UpperCamelCase_ ( self ) -> str: """simple docstring""" return (3, 32, 32) @property def UpperCamelCase_ ( self ) -> str: """simple docstring""" return (3, 32, 32) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = { '''block_out_channels''': (32, 64), '''down_block_types''': ('''DownBlock2D''', '''AttnDownBlock2D'''), '''up_block_types''': ('''AttnUpBlock2D''', '''UpBlock2D'''), '''attention_head_dim''': 3, '''out_channels''': 3, '''in_channels''': 3, '''layers_per_block''': 2, '''sample_size''': 32, } _lowerCamelCase = self.dummy_input return init_dict, inputs_dict class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = UNetaDModel A_ = 'sample' @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = 4 _lowerCamelCase = 4 _lowerCamelCase = (32, 32) _lowerCamelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) _lowerCamelCase = torch.tensor([10] ).to(A_ ) return {"sample": noise, "timestep": time_step} @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return (4, 32, 32) @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return (4, 32, 32) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = { '''sample_size''': 32, '''in_channels''': 4, '''out_channels''': 4, '''layers_per_block''': 2, '''block_out_channels''': (32, 64), '''attention_head_dim''': 32, '''down_block_types''': ('''DownBlock2D''', '''DownBlock2D'''), '''up_block_types''': ('''UpBlock2D''', '''UpBlock2D'''), } _lowerCamelCase = self.dummy_input return init_dict, inputs_dict def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase = UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=A_ ) self.assertIsNotNone(A_ ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(A_ ) _lowerCamelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != '''cuda''' , '''This test is supposed to run on GPU''' ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase = UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=A_ ) model.to(A_ ) _lowerCamelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != '''cuda''' , '''This test is supposed to run on GPU''' ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` _lowerCamelCase , _lowerCamelCase = UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' , output_loading_info=A_ ) model_accelerate.to(A_ ) model_accelerate.eval() _lowerCamelCase = torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) _lowerCamelCase = noise.to(A_ ) _lowerCamelCase = torch.tensor([10] * noise.shape[0] ).to(A_ ) _lowerCamelCase = model_accelerate(A_ , A_ )['''sample'''] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() _lowerCamelCase , _lowerCamelCase = UNetaDModel.from_pretrained( '''fusing/unet-ldm-dummy-update''' , output_loading_info=A_ , low_cpu_mem_usage=A_ ) model_normal_load.to(A_ ) model_normal_load.eval() _lowerCamelCase = model_normal_load(A_ , A_ )['''sample'''] assert torch_all_close(A_ , A_ , rtol=1E-3 ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = UNetaDModel.from_pretrained('''fusing/unet-ldm-dummy-update''' ) model.eval() model.to(A_ ) _lowerCamelCase = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) _lowerCamelCase = noise.to(A_ ) _lowerCamelCase = torch.tensor([10] * noise.shape[0] ).to(A_ ) with torch.no_grad(): _lowerCamelCase = model(A_ , A_ ).sample _lowerCamelCase = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off _lowerCamelCase = torch.tensor([-13.3258, -20.1100, -15.9873, -17.6617, -23.0596, -17.9419, -13.3675, -16.1889, -12.3800] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1E-3 ) ) class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = UNetaDModel A_ = 'sample' @property def UpperCamelCase_ ( self , A_=(32, 32) ) -> Optional[int]: """simple docstring""" _lowerCamelCase = 4 _lowerCamelCase = 3 _lowerCamelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) _lowerCamelCase = torch.tensor(batch_size * [10] ).to(dtype=torch.intaa , device=A_ ) return {"sample": noise, "timestep": time_step} @property def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" return (3, 32, 32) @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return (3, 32, 32) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = { '''block_out_channels''': [32, 64, 64, 64], '''in_channels''': 3, '''layers_per_block''': 1, '''out_channels''': 3, '''time_embedding_type''': '''fourier''', '''norm_eps''': 1E-6, '''mid_block_scale_factor''': math.sqrt(2.0 ), '''norm_num_groups''': None, '''down_block_types''': [ '''SkipDownBlock2D''', '''AttnSkipDownBlock2D''', '''SkipDownBlock2D''', '''SkipDownBlock2D''', ], '''up_block_types''': [ '''SkipUpBlock2D''', '''SkipUpBlock2D''', '''AttnSkipUpBlock2D''', '''SkipUpBlock2D''', ], } _lowerCamelCase = self.dummy_input return init_dict, inputs_dict @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase = UNetaDModel.from_pretrained('''google/ncsnpp-celebahq-256''' , output_loading_info=A_ ) self.assertIsNotNone(A_ ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(A_ ) _lowerCamelCase = self.dummy_input _lowerCamelCase = floats_tensor((4, 3) + (2_56, 2_56) ).to(A_ ) _lowerCamelCase = noise _lowerCamelCase = model(**A_ ) assert image is not None, "Make sure output is not None" @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = UNetaDModel.from_pretrained('''google/ncsnpp-celebahq-256''' ) model.to(A_ ) _lowerCamelCase = 4 _lowerCamelCase = 3 _lowerCamelCase = (2_56, 2_56) _lowerCamelCase = torch.ones((batch_size, num_channels) + sizes ).to(A_ ) _lowerCamelCase = torch.tensor(batch_size * [1E-4] ).to(A_ ) with torch.no_grad(): _lowerCamelCase = model(A_ , A_ ).sample _lowerCamelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off _lowerCamelCase = torch.tensor([-4842.8691, -6499.6631, -3800.1953, -7978.2686, -10980.7129, -20028.8535, 8148.2822, 2342.2905, 567.7608] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1E-2 ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = UNetaDModel.from_pretrained('''fusing/ncsnpp-ffhq-ve-dummy-update''' ) model.to(A_ ) _lowerCamelCase = 4 _lowerCamelCase = 3 _lowerCamelCase = (32, 32) _lowerCamelCase = torch.ones((batch_size, num_channels) + sizes ).to(A_ ) _lowerCamelCase = torch.tensor(batch_size * [1E-4] ).to(A_ ) with torch.no_grad(): _lowerCamelCase = model(A_ , A_ ).sample _lowerCamelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off _lowerCamelCase = torch.tensor([-0.0325, -0.0900, -0.0869, -0.0332, -0.0725, -0.0270, -0.0101, 0.0227, 0.0256] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1E-2 ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # not required for this model pass
638
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> None: """simple docstring""" warnings.warn( '''The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DPTImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
638
1
def __magic_name__( __UpperCAmelCase = 200_0000 ) -> int: '''simple docstring''' _lowerCamelCase = [0 for i in range(n + 1 )] _lowerCamelCase = 1 _lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , __UpperCAmelCase ): _lowerCamelCase = 1 _lowerCamelCase = 0 for i in range(__UpperCAmelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f'''{solution() = }''')
638
import argparse import json import subprocess def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = ( F'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) _lowerCamelCase = subprocess.run(__UpperCAmelCase , shell=__UpperCAmelCase , stdout=subprocess.PIPE ) _lowerCamelCase = output.stdout.decode('''utf-8''' ) _lowerCamelCase = json.loads(__UpperCAmelCase ) _lowerCamelCase = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__UpperCAmelCase ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) ) if len(__UpperCAmelCase ) > 0: _lowerCamelCase = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' return values.split(''',''' ) snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) snake_case__ = parser.parse_args() get_runner_status(args.target_runners, args.token)
638
1
snake_case__ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> list[str]: '''simple docstring''' _lowerCamelCase = set() # keep track of all the paths to be checked _lowerCamelCase = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue _lowerCamelCase = queue.pop(0 ) # get the last node from the path _lowerCamelCase = path[-1] if node not in explored: _lowerCamelCase = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: _lowerCamelCase = list(__UpperCAmelCase ) new_path.append(__UpperCAmelCase ) queue.append(__UpperCAmelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(__UpperCAmelCase ) # in case there's no path between the 2 nodes return [] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 _lowerCamelCase = [start] _lowerCamelCase = set(__UpperCAmelCase ) # Keep tab on distances from `start` node. _lowerCamelCase = {start: 0, target: -1} while queue: _lowerCamelCase = queue.pop(0 ) if node == target: _lowerCamelCase = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(__UpperCAmelCase ) queue.append(__UpperCAmelCase ) _lowerCamelCase = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
638
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
638
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase snake_case__ = logging.get_logger(__name__) snake_case__ = { 'allenai/longformer-base-4096': 'https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json', 'allenai/longformer-large-4096': 'https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json', 'allenai/longformer-large-4096-finetuned-triviaqa': ( 'https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json' ), 'allenai/longformer-base-4096-extra.pos.embd.only': ( 'https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json' ), 'allenai/longformer-large-4096-extra.pos.embd.only': ( 'https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json' ), } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'longformer' def __init__( self , A_ = 5_12 , A_ = 2 , A_ = 1 , A_ = 0 , A_ = 2 , A_ = 3_05_22 , A_ = 7_68 , A_ = 12 , A_ = 12 , A_ = 30_72 , A_ = "gelu" , A_ = 0.1 , A_ = 0.1 , A_ = 5_12 , A_ = 2 , A_ = 0.02 , A_ = 1E-1_2 , A_ = False , **A_ , ) -> int: """simple docstring""" super().__init__(pad_token_id=A_ , **A_ ) _lowerCamelCase = attention_window _lowerCamelCase = sep_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = eos_token_id _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = onnx_export class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , A_ , A_ = "default" , A_ = None ) -> List[Any]: """simple docstring""" super().__init__(A_ , A_ , A_ ) _lowerCamelCase = True @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": _lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''global_attention_mask''', dynamic_axis), ] ) @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" _lowerCamelCase = super().outputs if self.task == "default": _lowerCamelCase = {0: '''batch'''} return outputs @property def UpperCamelCase_ ( self ) -> float: """simple docstring""" return 1E-4 @property def UpperCamelCase_ ( self ) -> int: """simple docstring""" # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def UpperCamelCase_ ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) -> Mapping[str, Any]: """simple docstring""" _lowerCamelCase = super().generate_dummy_inputs( preprocessor=A_ , batch_size=A_ , seq_length=A_ , is_pair=A_ , framework=A_ ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly _lowerCamelCase = torch.zeros_like(inputs['''input_ids'''] ) # make every second token global _lowerCamelCase = 1 return inputs
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
1
import datasets from .evaluate import evaluate snake_case__ = '\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n' snake_case__ = '\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n' snake_case__ = '\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the CUAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\n \'aupr\': Area Under the Precision-Recall curve\n \'prec_at_80_recall\': Precision at 80% recall\n \'prec_at_90_recall\': Precision at 90% recall\nExamples:\n >>> predictions = [{\'prediction_text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\'], \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> references = [{\'answers\': {\'answer_start\': [143, 49], \'text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\']}, \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> cuad_metric = datasets.load_metric("cuad")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0, \'aupr\': 0.0, \'prec_at_80_recall\': 1.0, \'prec_at_90_recall\': 1.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': { '''id''': datasets.Value('''string''' ), '''prediction_text''': datasets.features.Sequence(datasets.Value('''string''' ) ), }, '''references''': { '''id''': datasets.Value('''string''' ), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), }, } ) , codebase_urls=['''https://www.atticusprojectai.org/cuad'''] , reference_urls=['''https://www.atticusprojectai.org/cuad'''] , ) def UpperCamelCase_ ( self , A_ , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} _lowerCamelCase = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] _lowerCamelCase = evaluate(dataset=A_ , predictions=A_ ) return score
638
import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaTokenizerFast, ) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = model.config _lowerCamelCase = DonutSwinConfig( image_size=original_config.input_size , patch_size=4 , depths=original_config.encoder_layer , num_heads=[4, 8, 16, 32] , window_size=original_config.window_size , embed_dim=128 , ) _lowerCamelCase = MBartConfig( is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , add_cross_attention=__UpperCAmelCase , decoder_layers=original_config.decoder_layer , max_position_embeddings=original_config.max_position_embeddings , vocab_size=len( model.decoder.tokenizer ) , scale_embedding=__UpperCAmelCase , add_final_layer_norm=__UpperCAmelCase , ) return encoder_config, decoder_config def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' if "encoder.model" in name: _lowerCamelCase = name.replace('''encoder.model''' , '''encoder''' ) if "decoder.model" in name: _lowerCamelCase = name.replace('''decoder.model''' , '''decoder''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if name.startswith('''encoder''' ): if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "mask" not in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": _lowerCamelCase = '''encoder.layernorm.weight''' if name == "encoder.norm.bias": _lowerCamelCase = '''encoder.layernorm.bias''' return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = int(key_split[5] ) _lowerCamelCase = model.encoder.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] elif "attn_mask" in key or key in ["encoder.model.norm.weight", "encoder.model.norm.bias"]: # HuggingFace implementation doesn't use attn_mask buffer # and model doesn't use final LayerNorms for the encoder pass else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = DonutModel.from_pretrained(__UpperCAmelCase ).eval() # load HuggingFace model _lowerCamelCase , _lowerCamelCase = get_configs(__UpperCAmelCase ) _lowerCamelCase = DonutSwinModel(__UpperCAmelCase ) _lowerCamelCase = MBartForCausalLM(__UpperCAmelCase ) _lowerCamelCase = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() _lowerCamelCase = original_model.state_dict() _lowerCamelCase = convert_state_dict(__UpperCAmelCase , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) # verify results on scanned document _lowerCamelCase = load_dataset('''hf-internal-testing/example-documents''' ) _lowerCamelCase = dataset['''test'''][0]['''image'''].convert('''RGB''' ) _lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained(__UpperCAmelCase , from_slow=__UpperCAmelCase ) _lowerCamelCase = DonutImageProcessor( do_align_long_axis=original_model.config.align_long_axis , size=original_model.config.input_size[::-1] ) _lowerCamelCase = DonutProcessor(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values if model_name == "naver-clova-ix/donut-base-finetuned-docvqa": _lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' _lowerCamelCase = '''When is the coffee break?''' _lowerCamelCase = task_prompt.replace('''{user_input}''' , __UpperCAmelCase ) elif model_name == "naver-clova-ix/donut-base-finetuned-rvlcdip": _lowerCamelCase = '''<s_rvlcdip>''' elif model_name in [ "naver-clova-ix/donut-base-finetuned-cord-v1", "naver-clova-ix/donut-base-finetuned-cord-v1-2560", ]: _lowerCamelCase = '''<s_cord>''' elif model_name == "naver-clova-ix/donut-base-finetuned-cord-v2": _lowerCamelCase = '''s_cord-v2>''' elif model_name == "naver-clova-ix/donut-base-finetuned-zhtrainticket": _lowerCamelCase = '''<s_zhtrainticket>''' elif model_name in ["naver-clova-ix/donut-proto", "naver-clova-ix/donut-base"]: # use a random prompt _lowerCamelCase = '''hello world''' else: raise ValueError('''Model name not supported''' ) _lowerCamelCase = original_model.decoder.tokenizer(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors='''pt''' )[ '''input_ids''' ] _lowerCamelCase = original_model.encoder.model.patch_embed(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = model.encoder.embeddings(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) # verify encoder hidden states _lowerCamelCase = original_model.encoder(__UpperCAmelCase ) _lowerCamelCase = model.encoder(__UpperCAmelCase ).last_hidden_state assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) # verify decoder hidden states _lowerCamelCase = original_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ).logits _lowerCamelCase = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: model.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) processor.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='naver-clova-ix/donut-base-finetuned-docvqa', required=False, type=str, help='Name of the original model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, required=False, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub.', ) snake_case__ = parser.parse_args() convert_donut_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
1
import os from distutils.util import strtobool def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' for e in env_keys: _lowerCamelCase = int(os.environ.get(__UpperCAmelCase , -1 ) ) if val >= 0: return val return default def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> Any: '''simple docstring''' _lowerCamelCase = os.environ.get(__UpperCAmelCase , str(__UpperCAmelCase ) ) return strtobool(__UpperCAmelCase ) == 1 # As its name indicates `strtobool` actually returns an int... def __magic_name__( __UpperCAmelCase , __UpperCAmelCase="no" ) -> int: '''simple docstring''' _lowerCamelCase = os.environ.get(__UpperCAmelCase , str(__UpperCAmelCase ) ) return value
638
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__UpperCAmelCase ) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' if index == len(__UpperCAmelCase ): return True # Recursive Step for i in range(__UpperCAmelCase ): if valid_coloring(graph[index] , __UpperCAmelCase , __UpperCAmelCase ): # Color current vertex _lowerCamelCase = i # Validate coloring if util_color(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , index + 1 ): return True # Backtrack _lowerCamelCase = -1 return False def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> list[int]: '''simple docstring''' _lowerCamelCase = [-1] * len(__UpperCAmelCase ) if util_color(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , 0 ): return colored_vertices return []
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
def __magic_name__( __UpperCAmelCase = 3 , __UpperCAmelCase = 7 , __UpperCAmelCase = 100_0000 ) -> int: '''simple docstring''' _lowerCamelCase = 0 _lowerCamelCase = 1 for current_denominator in range(1 , limit + 1 ): _lowerCamelCase = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: _lowerCamelCase = current_numerator _lowerCamelCase = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=100_0000))
638
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
1
import re from filelock import FileLock try: import nltk snake_case__ = True except (ImportError, ModuleNotFoundError): snake_case__ = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' re.sub('''<n>''' , '''''' , __UpperCAmelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__UpperCAmelCase ) )
638
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class UpperCamelCase ( __lowercase ): '''simple docstring''' @staticmethod @abstractmethod def UpperCamelCase_ ( A_ ) -> List[str]: """simple docstring""" raise NotImplementedError() @abstractmethod def UpperCamelCase_ ( self ) -> Any: """simple docstring""" raise NotImplementedError()
638
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = (KDPMaDiscreteScheduler,) A_ = 10 def UpperCamelCase_ ( self , **A_ ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = { '''num_train_timesteps''': 11_00, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**A_ ) return config def UpperCamelCase_ ( self ) -> int: """simple docstring""" for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=A_ , beta_end=A_ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config(prediction_type='''v_prediction''' ) _lowerCamelCase = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma _lowerCamelCase = sample.to(A_ ) for i, t in enumerate(scheduler.timesteps ): _lowerCamelCase = scheduler.scale_model_input(A_ , A_ ) _lowerCamelCase = model(A_ , A_ ) _lowerCamelCase = scheduler.step(A_ , A_ , A_ ) _lowerCamelCase = output.prev_sample _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_9_3_4E-0_7 ) < 1E-2 assert abs(result_mean.item() - 6.1_1_1_2E-1_0 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.6_9_3_4_2_8_6_5_0_1_7_0_9_7_2E-0_7 ) < 1E-2 assert abs(result_mean.item() - 0.0002 ) < 1E-3 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" if torch_device == "mps": return _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma _lowerCamelCase = sample.to(A_ ) for i, t in enumerate(scheduler.timesteps ): _lowerCamelCase = scheduler.scale_model_input(A_ , A_ ) _lowerCamelCase = model(A_ , A_ ) _lowerCamelCase = scheduler.step(A_ , A_ , A_ ) _lowerCamelCase = output.prev_sample _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" if torch_device == "mps": return _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps , device=A_ ) _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter.to(A_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: _lowerCamelCase = scheduler.scale_model_input(A_ , A_ ) _lowerCamelCase = model(A_ , A_ ) _lowerCamelCase = scheduler.step(A_ , A_ , A_ ) _lowerCamelCase = output.prev_sample _lowerCamelCase = torch.sum(torch.abs(A_ ) ) _lowerCamelCase = torch.mean(torch.abs(A_ ) ) if str(A_ ).startswith('''cpu''' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3
638
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } _lowerCamelCase = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub _lowerCamelCase = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase_ ( self , **A_ ) -> str: """simple docstring""" _lowerCamelCase = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> Optional[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> int: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = feature_extractor(A_ , return_tensors='''np''' ) _lowerCamelCase = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = '''This is a test string''' _lowerCamelCase = processor(text=A_ ) _lowerCamelCase = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase_ ( self , A_=(2, 10, 16) , A_=77 ) -> Optional[Any]: """simple docstring""" np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits(shape=(10, 16) , seed=13 ) _lowerCamelCase = processor.decode(A_ ) _lowerCamelCase = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: _lowerCamelCase = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: _lowerCamelCase = processor.batch_decode(A_ , A_ ) _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as p: _lowerCamelCase = decoder.decode_beams_batch(A_ , A_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 15 _lowerCamelCase = -20.0 _lowerCamelCase = -4.0 _lowerCamelCase = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] _lowerCamelCase = [d[0][2] for d in decoded_decoder_out] _lowerCamelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 2.0 _lowerCamelCase = 5.0 _lowerCamelCase = -20.0 _lowerCamelCase = True _lowerCamelCase = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = snapshot_download('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = processor_wavaveca(A_ , return_tensors='''np''' ) _lowerCamelCase = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor_wavaveca.batch_decode(A_ ) _lowerCamelCase = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = [d[key] for d in offsets] return retrieved_list def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits()[0] _lowerCamelCase = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" import torch _lowerCamelCase = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) _lowerCamelCase = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) _lowerCamelCase = iter(A_ ) _lowerCamelCase = next(A_ ) _lowerCamelCase = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) _lowerCamelCase = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train _lowerCamelCase = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): _lowerCamelCase = model(A_ ).logits.cpu().numpy() _lowerCamelCase = processor.decode(logits[0] , output_word_offsets=A_ ) _lowerCamelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate _lowerCamelCase = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] _lowerCamelCase = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off _lowerCamelCase = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) _lowerCamelCase = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
638
1
import warnings warnings.warn( 'memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ' '`from accelerate import find_executable_batch_size` to avoid this warning.', FutureWarning, )
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _lowerCamelCase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _lowerCamelCase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _lowerCamelCase = subset[i - 1][j] if arr[i - 1] <= j: _lowerCamelCase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
638
1
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class UpperCamelCase : '''simple docstring''' def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=4 , A_="gelu" , A_=0.0 , A_=0.1 , A_=True , A_=5_12 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ) -> Dict: """simple docstring""" _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_input_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_multiple_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout _lowerCamelCase = attention_dropout _lowerCamelCase = weight_tying _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self ) -> Any: """simple docstring""" return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase = True return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = GPTNeoXJapaneseModel(config=A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ ) _lowerCamelCase = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = True _lowerCamelCase = GPTNeoXJapaneseModel(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> Tuple: """simple docstring""" _lowerCamelCase = GPTNeoXJapaneseForCausalLM(config=A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = True _lowerCamelCase = GPTNeoXJapaneseForCausalLM(config=A_ ) model.to(A_ ) model.eval() # first forward pass _lowerCamelCase = model(A_ , attention_mask=A_ , use_cache=A_ ) _lowerCamelCase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCamelCase = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = torch.cat([input_mask, next_mask] , dim=-1 ) _lowerCamelCase = model(A_ , attention_mask=A_ , output_hidden_states=A_ ) _lowerCamelCase = output_from_no_past['''hidden_states'''][0] _lowerCamelCase = model( A_ , attention_mask=A_ , past_key_values=A_ , output_hidden_states=A_ , )['''hidden_states'''][0] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A_ , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () A_ = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () A_ = ( {'feature-extraction': GPTNeoXJapaneseModel, 'text-generation': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) A_ = False A_ = False A_ = False A_ = False def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = GPTNeoXJapaneseModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=A_ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" # This regression test was failing with PyTorch < 1.3 _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() _lowerCamelCase = None self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*A_ ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''abeja/gpt-neox-japanese-2.7b''' _lowerCamelCase = ['''データサイエンティストとは、''', '''100年後に必要とされる会社は、''', '''フルリモートの環境で働くために必要なことは、''', '''国境の長いトンネルを抜けると''', '''美味しい日本食といえば、'''] _lowerCamelCase = [ '''データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。''', '''100年後に必要とされる会社は、「人」が中心の会社です。''', '''フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。''', '''国境の長いトンネルを抜けると、そこは雪国だった。''', '''美味しい日本食といえば、やっぱりお寿司ですよね。''', ] _lowerCamelCase = GPTNeoXJapaneseTokenizer.from_pretrained(A_ ) _lowerCamelCase = GPTNeoXJapaneseForCausalLM.from_pretrained(A_ ) _lowerCamelCase = [] for prompt in prompts: _lowerCamelCase = tokenizer(A_ , return_tensors='''pt''' ).input_ids _lowerCamelCase = model.generate(A_ , max_length=50 ) _lowerCamelCase = tokenizer.batch_decode(A_ , skip_special_tokens=A_ ) predicted_outputs += generated_string self.assertListEqual(A_ , A_ )
638
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
1
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def __magic_name__( __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = np.max(_outputs , axis=-1 , keepdims=__UpperCAmelCase ) _lowerCamelCase = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=__UpperCAmelCase ) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'sigmoid' A_ = 'softmax' A_ = 'none' @add_end_docstrings( __lowercase , R'\n return_all_scores (`bool`, *optional*, defaults to `False`):\n Whether to return all prediction scores or just the one of the predicted class.\n function_to_apply (`str`, *optional*, defaults to `"default"`):\n The function to apply to the model outputs in order to retrieve the scores. Accepts four different values:\n\n - `"default"`: if the model has a single label, will apply the sigmoid function on the output. If the model\n has several labels, will apply the softmax function on the output.\n - `"sigmoid"`: Applies the sigmoid function on the output.\n - `"softmax"`: Applies the softmax function on the output.\n - `"none"`: Does not apply any function on the output.\n ' , ) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = False A_ = ClassificationFunction.NONE def __init__( self , **A_ ) -> str: """simple docstring""" super().__init__(**A_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def UpperCamelCase_ ( self , A_=None , A_=None , A_="" , **A_ ) -> str: """simple docstring""" # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" _lowerCamelCase = tokenizer_kwargs _lowerCamelCase = {} if hasattr(self.model.config , '''return_all_scores''' ) and return_all_scores is None: _lowerCamelCase = self.model.config.return_all_scores if isinstance(A_ , A_ ) or top_k is None: _lowerCamelCase = top_k _lowerCamelCase = False elif return_all_scores is not None: warnings.warn( '''`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of''' ''' `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.''' , A_ , ) if return_all_scores: _lowerCamelCase = None else: _lowerCamelCase = 1 if isinstance(A_ , A_ ): _lowerCamelCase = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: _lowerCamelCase = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *A_ , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = super().__call__(*A_ , **A_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. _lowerCamelCase = '''top_k''' not in kwargs if isinstance(args[0] , A_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def UpperCamelCase_ ( self , A_ , **A_ ) -> Dict[str, GenericTensor]: """simple docstring""" _lowerCamelCase = self.framework if isinstance(A_ , A_ ): return self.tokenizer(**A_ , return_tensors=A_ , **A_ ) elif isinstance(A_ , A_ ) and len(A_ ) == 1 and isinstance(inputs[0] , A_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=A_ , **A_ ) elif isinstance(A_ , A_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( '''The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a''' ''' dictionary `{"text": "My text", "text_pair": "My pair"}` in order to send a text pair.''' ) return self.tokenizer(A_ , return_tensors=A_ , **A_ ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" return self.model(**A_ ) def UpperCamelCase_ ( self , A_ , A_=None , A_=1 , A_=True ) -> List[Any]: """simple docstring""" # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: _lowerCamelCase = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: _lowerCamelCase = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , '''function_to_apply''' ) and function_to_apply is None: _lowerCamelCase = self.model.config.function_to_apply else: _lowerCamelCase = ClassificationFunction.NONE _lowerCamelCase = model_outputs['''logits'''][0] _lowerCamelCase = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: _lowerCamelCase = sigmoid(A_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: _lowerCamelCase = softmax(A_ ) elif function_to_apply == ClassificationFunction.NONE: _lowerCamelCase = outputs else: raise ValueError(F'Unrecognized `function_to_apply` argument: {function_to_apply}' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} _lowerCamelCase = [ {'''label''': self.model.config.idalabel[i], '''score''': score.item()} for i, score in enumerate(A_ ) ] if not _legacy: dict_scores.sort(key=lambda A_ : x["score"] , reverse=A_ ) if top_k is not None: _lowerCamelCase = dict_scores[:top_k] return dict_scores
638
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
1
snake_case__ = { "joule": 1.0, "kilojoule": 1000, "megajoule": 100_0000, "gigajoule": 10_0000_0000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 360_0000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 418_6800.00, "electronvolt": 1.6_0217_6634E-19, "britishthermalunit_it": 1055.0_5585, "footpound": 1.355818, } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> float: '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: _lowerCamelCase = ( F'Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n' F'Valid values are: {", ".join(__UpperCAmelCase )}' ) raise ValueError(__UpperCAmelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
638
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger snake_case__ = get_logger(__name__) class UpperCamelCase ( enum.Enum ): '''simple docstring''' A_ = 'all_checks' A_ = 'basic_checks' A_ = 'no_checks' class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ) -> Tuple: '''simple docstring''' if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) ) if len(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) ) _lowerCamelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _lowerCamelCase = ''' for ''' + verification_name if verification_name is not None else '''''' if len(__UpperCAmelCase ) > 0: raise NonMatchingChecksumError( F'Checksums didn\'t match{for_verification_name}:\n' F'{bad_urls}\n' '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' class UpperCamelCase ( __lowercase ): '''simple docstring''' def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) ) if len(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__UpperCAmelCase ) - set(__UpperCAmelCase ) ) ) _lowerCamelCase = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__UpperCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__UpperCAmelCase ) ) logger.info('''All the splits matched successfully.''' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = True ) -> dict: '''simple docstring''' if record_checksum: _lowerCamelCase = shaaaa() with open(__UpperCAmelCase , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'''''' ): m.update(__UpperCAmelCase ) _lowerCamelCase = m.hexdigest() else: _lowerCamelCase = None return {"num_bytes": os.path.getsize(__UpperCAmelCase ), "checksum": checksum} def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
638
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ) -> None: """simple docstring""" _lowerCamelCase = device _lowerCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _lowerCamelCase = [0.48145466, 0.4578275, 0.40821073] _lowerCamelCase = [0.26862954, 0.26130258, 0.27577711] _lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCamelCase = torchvision.transforms.Resize(2_24 ) _lowerCamelCase = torchvision.transforms.CenterCrop(2_24 ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.resize(A_ ) _lowerCamelCase = self.center_crop(A_ ) _lowerCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer(text=A_ , **A_ ) _lowerCamelCase = self.preprocess_img(A_ ) _lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ) -> None: """simple docstring""" super().__init__() _lowerCamelCase = None _lowerCamelCase = device if device else get_device() if vqgan: _lowerCamelCase = vqgan else: _lowerCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _lowerCamelCase = clip else: _lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _lowerCamelCase = ProcessorGradientFlow(device=self.device ) _lowerCamelCase = iterations _lowerCamelCase = lr _lowerCamelCase = log _lowerCamelCase = make_grid _lowerCamelCase = return_val _lowerCamelCase = quantize _lowerCamelCase = self.vqgan.decoder.z_shape def UpperCamelCase_ ( self , A_=None , A_=None , A_=5 , A_=True ) -> Any: """simple docstring""" _lowerCamelCase = [] if output_path is None: _lowerCamelCase = '''./animation.gif''' if input_path is None: _lowerCamelCase = self.save_path _lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(A_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(A_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _lowerCamelCase = total_duration / len(A_ ) _lowerCamelCase = [frame_duration] * len(A_ ) if extend_frames: _lowerCamelCase = 1.5 _lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F'gif saved to {output_path}' ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> Union[str, Any]: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _lowerCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _lowerCamelCase = preprocess_vqgan(A_ ) _lowerCamelCase , *_lowerCamelCase = self.vqgan.encode(A_ ) return z def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.latent.detach().requires_grad_() _lowerCamelCase = base_latent + transform_vector if self.quantize: _lowerCamelCase , *_lowerCamelCase = self.vqgan.quantize(A_ ) else: _lowerCamelCase = trans_latent return self.vqgan.decode(A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_=None ) -> Any: """simple docstring""" _lowerCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors='''pt''' , padding=A_ ) _lowerCamelCase = self.clip(**A_ ) _lowerCamelCase = clip_outputs.logits_per_image if weights is not None: _lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , A_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , A_ , weights=neg_prompts['''weights'''] ) else: _lowerCamelCase = torch.tensor([1] , device=self.device ) _lowerCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCamelCase = self._add_vector(A_ ) _lowerCamelCase = loop_post_process(A_ ) _lowerCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print('''CLIP loss''' , A_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" wandb.init(reinit=A_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _lowerCamelCase = Image.open(A_ ) _lowerCamelCase = image.resize((2_56, 2_56) ) wandb.log('''Original Image''' , wandb.Image(A_ ) ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if not prompts: return [] _lowerCamelCase = [] _lowerCamelCase = [] if isinstance(A_ , A_ ): _lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _lowerCamelCase = prompt[0] _lowerCamelCase = float(prompt[1] ) elif ":" in prompt: _lowerCamelCase , _lowerCamelCase = prompt.split(''':''' ) _lowerCamelCase = float(A_ ) else: _lowerCamelCase = prompt _lowerCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ) -> str: """simple docstring""" if image_path: _lowerCamelCase = self._get_latent(A_ ) else: _lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCamelCase = self.process_prompts(A_ ) _lowerCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(A_ ) _lowerCamelCase = save_path _lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(A_ ) ) _lowerCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({'''Image''': wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
638
1
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = RemBertConfig.from_json_file(__UpperCAmelCase ) print('''Building PyTorch model from configuration: {}'''.format(str(__UpperCAmelCase ) ) ) _lowerCamelCase = RemBertModel(__UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_rembert(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # Save pytorch-model print('''Save PyTorch model to {}'''.format(__UpperCAmelCase ) ) torch.save(model.state_dict() , __UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--rembert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained RemBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) snake_case__ = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
638
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case__ = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) snake_case__ = { 'iou_prediction_head.layers.0': 'iou_prediction_head.proj_in', 'iou_prediction_head.layers.1': 'iou_prediction_head.layers.0', 'iou_prediction_head.layers.2': 'iou_prediction_head.proj_out', 'mask_decoder.output_upscaling.0': 'mask_decoder.upscale_conv1', 'mask_decoder.output_upscaling.1': 'mask_decoder.upscale_layer_norm', 'mask_decoder.output_upscaling.3': 'mask_decoder.upscale_conv2', 'mask_downscaling.0': 'mask_embed.conv1', 'mask_downscaling.1': 'mask_embed.layer_norm1', 'mask_downscaling.3': 'mask_embed.conv2', 'mask_downscaling.4': 'mask_embed.layer_norm2', 'mask_downscaling.6': 'mask_embed.conv3', 'point_embeddings': 'point_embed', 'pe_layer.positional_encoding_gaussian_matrix': 'shared_embedding.positional_embedding', 'image_encoder': 'vision_encoder', 'neck.0': 'neck.conv1', 'neck.1': 'neck.layer_norm1', 'neck.2': 'neck.conv2', 'neck.3': 'neck.layer_norm2', 'patch_embed.proj': 'patch_embed.projection', '.norm': '.layer_norm', 'blocks': 'layers', } def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' _lowerCamelCase = {} state_dict.pop('''pixel_mean''' , __UpperCAmelCase ) state_dict.pop('''pixel_std''' , __UpperCAmelCase ) _lowerCamelCase = r'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _lowerCamelCase = key.replace(__UpperCAmelCase , __UpperCAmelCase ) if re.match(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = int(re.match(__UpperCAmelCase , __UpperCAmelCase ).group(2 ) ) if layer_nb == 0: _lowerCamelCase = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: _lowerCamelCase = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: _lowerCamelCase = key.replace('''layers.2''' , '''proj_out''' ) _lowerCamelCase = value _lowerCamelCase = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase="ybelkada/segment-anything" ) -> Any: '''simple docstring''' _lowerCamelCase = hf_hub_download(__UpperCAmelCase , F'checkpoints/{model_name}.pth' ) if "sam_vit_b" in model_name: _lowerCamelCase = SamConfig() elif "sam_vit_l" in model_name: _lowerCamelCase = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) _lowerCamelCase = SamConfig( vision_config=__UpperCAmelCase , ) elif "sam_vit_h" in model_name: _lowerCamelCase = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) _lowerCamelCase = SamConfig( vision_config=__UpperCAmelCase , ) _lowerCamelCase = torch.load(__UpperCAmelCase , map_location='''cpu''' ) _lowerCamelCase = replace_keys(__UpperCAmelCase ) _lowerCamelCase = SamImageProcessor() _lowerCamelCase = SamProcessor(image_processor=__UpperCAmelCase ) _lowerCamelCase = SamModel(__UpperCAmelCase ) hf_model.load_state_dict(__UpperCAmelCase ) _lowerCamelCase = hf_model.to('''cuda''' ) _lowerCamelCase = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ).convert('''RGB''' ) _lowerCamelCase = [[[400, 650]]] _lowerCamelCase = [[1]] _lowerCamelCase = processor(images=np.array(__UpperCAmelCase ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**__UpperCAmelCase ) _lowerCamelCase = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.5_7_9_8_9_0_2_5_1_1_5_9_6_6_8 _lowerCamelCase = processor( images=np.array(__UpperCAmelCase ) , input_points=__UpperCAmelCase , input_labels=__UpperCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**__UpperCAmelCase ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_7_1_2_6_0_3_0_9_2_1_9_3_6_0_4 _lowerCamelCase = ((75, 275, 1725, 850),) _lowerCamelCase = processor(images=np.array(__UpperCAmelCase ) , input_boxes=__UpperCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**__UpperCAmelCase ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_6_8_6_0_1_5_6_0_5_9_2_6_5_1_4 # Test with 2 points and 1 image. _lowerCamelCase = [[[400, 650], [800, 650]]] _lowerCamelCase = [[1, 1]] _lowerCamelCase = processor( images=np.array(__UpperCAmelCase ) , input_points=__UpperCAmelCase , input_labels=__UpperCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**__UpperCAmelCase ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_9_3_6_0_4_7_7_9_2_4_3_4_6_9_2 if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() snake_case__ = ['sam_vit_b_01ec64', 'sam_vit_h_4b8939', 'sam_vit_l_0b3195'] parser.add_argument( '--model_name', default='sam_vit_h_4b8939', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) parser.add_argument( '--model_hub_id', default='ybelkada/segment-anything', choices=choices, type=str, help='Path to hf config.json of model to convert', ) snake_case__ = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
638
import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='%(message)s') def __magic_name__( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size, 1) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.mean(1 ) # Centralize the data of class i _lowerCamelCase = data - column_reshape(__UpperCAmelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(__UpperCAmelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = features.mean(1 ) _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.shape[1] _lowerCamelCase = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' if features.any(): _lowerCamelCase = features.mean(1 ) # Center the dataset _lowerCamelCase = features - np.reshape(__UpperCAmelCase , (data_mean.size, 1) ) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) / features.shape[1] _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(__UpperCAmelCase ) # Take all the columns in the reverse order (-1), and then takes only the first _lowerCamelCase = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space _lowerCamelCase = np.dot(filtered_eigenvectors.T , __UpperCAmelCase ) logging.info('''Principal Component Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' assert classes > dimensions # Check if features have been already loaded if features.any: _lowerCamelCase , _lowerCamelCase = eigh( covariance_between_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , covariance_within_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , ) _lowerCamelCase = eigenvectors[:, ::-1][:, :dimensions] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = np.linalg.svd(__UpperCAmelCase ) _lowerCamelCase = svd_matrix[:, 0:dimensions] _lowerCamelCase = np.dot(filtered_svd_matrix.T , __UpperCAmelCase ) logging.info('''Linear Discriminant Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) _lowerCamelCase = np.array([0, 0, 0, 1, 1] ) _lowerCamelCase = 2 _lowerCamelCase = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = linear_discriminant_analysis( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if isinstance(__UpperCAmelCase , np.ndarray ): raise AssertionError( '''Did not raise AssertionError for dimensions > classes''' ) assert error_info.type is AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) _lowerCamelCase = 2 _lowerCamelCase = np.array([[6.9_2_8_2_0_3_2_3, 8.6_6_0_2_5_4_0_4, 1_0.3_9_2_3_0_4_8_5], [3.0, 3.0, 3.0]] ) with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = principal_component_analysis(__UpperCAmelCase , __UpperCAmelCase ) if not np.allclose(__UpperCAmelCase , __UpperCAmelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
638
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['image_processor', 'tokenizer'] A_ = 'LayoutLMv2ImageProcessor' A_ = ('LayoutXLMTokenizer', 'LayoutXLMTokenizerFast') def __init__( self , A_=None , A_=None , **A_ ) -> Tuple: """simple docstring""" if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A_ , ) _lowerCamelCase = kwargs.pop('''feature_extractor''' ) _lowerCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(A_ , A_ ) def __call__( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ) -> BatchEncoding: """simple docstring""" # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''' ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''' ) # first, apply the image processor _lowerCamelCase = self.image_processor(images=A_ , return_tensors=A_ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(A_ , A_ ): _lowerCamelCase = [text] # add batch dimension (as the image processor always adds a batch dimension) _lowerCamelCase = features['''words'''] _lowerCamelCase = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_token_type_ids=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) # add pixel values _lowerCamelCase = features.pop('''pixel_values''' ) if return_overflowing_tokens is True: _lowerCamelCase = self.get_overflowing_images(A_ , encoded_inputs['''overflow_to_sample_mapping'''] ) _lowerCamelCase = images return encoded_inputs def UpperCamelCase_ ( self , A_ , A_ ) -> Tuple: """simple docstring""" # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image _lowerCamelCase = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(A_ ) != len(A_ ): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' F' {len(A_ )} and {len(A_ )}' ) return images_with_overflow def UpperCamelCase_ ( self , *A_ , **A_ ) -> Any: """simple docstring""" return self.tokenizer.batch_decode(*A_ , **A_ ) def UpperCamelCase_ ( self , *A_ , **A_ ) -> Any: """simple docstring""" return self.tokenizer.decode(*A_ , **A_ ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return ["input_ids", "bbox", "attention_mask", "image"] @property def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , A_ , ) return self.image_processor_class @property def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , A_ , ) return self.image_processor
638
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['vqvae'] def __init__( self , A_ , A_ , A_ , A_ , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules(unet=A_ , scheduler=A_ , mel=A_ , vqvae=A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 50 if isinstance(self.scheduler , A_ ) else 10_00 @torch.no_grad() def __call__( self , A_ = 1 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = 0 , A_ = None , A_ = None , A_=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: """simple docstring""" _lowerCamelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(A_ ) _lowerCamelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: _lowerCamelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: _lowerCamelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=A_ , device=self.device , ) _lowerCamelCase = noise _lowerCamelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A_ , A_ ) _lowerCamelCase = self.mel.audio_slice_to_image(A_ ) _lowerCamelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) _lowerCamelCase = (input_image / 2_55) * 2 - 1 _lowerCamelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: _lowerCamelCase = self.vqvae.encode(torch.unsqueeze(A_ , 0 ) ).latent_dist.sample( generator=A_ )[0] _lowerCamelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , self.scheduler.timesteps[start_step - 1] ) _lowerCamelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) _lowerCamelCase = int(mask_start_secs * pixels_per_second ) _lowerCamelCase = int(mask_end_secs * pixels_per_second ) _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , A_ ): _lowerCamelCase = self.unet(A_ , A_ , A_ )['''sample'''] else: _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] if isinstance(self.scheduler , A_ ): _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , eta=A_ , generator=A_ , )['''prev_sample'''] else: _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , generator=A_ , )['''prev_sample'''] if mask is not None: if mask_start > 0: _lowerCamelCase = mask[:, step, :, :mask_start] if mask_end > 0: _lowerCamelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance _lowerCamelCase = 1 / self.vqvae.config.scaling_factor * images _lowerCamelCase = self.vqvae.decode(A_ )['''sample'''] _lowerCamelCase = (images / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() _lowerCamelCase = (images * 2_55).round().astype('''uint8''' ) _lowerCamelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A_ , mode='''RGB''' ).convert('''L''' ) for _ in images) ) _lowerCamelCase = [self.mel.image_to_audio(A_ ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A_ )[:, np.newaxis, :] ) , **ImagePipelineOutput(A_ ) ) @torch.no_grad() def UpperCamelCase_ ( self , A_ , A_ = 50 ) -> np.ndarray: """simple docstring""" assert isinstance(self.scheduler , A_ ) self.scheduler.set_timesteps(A_ ) _lowerCamelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) _lowerCamelCase = (sample / 2_55) * 2 - 1 _lowerCamelCase = torch.Tensor(A_ ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): _lowerCamelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps _lowerCamelCase = self.scheduler.alphas_cumprod[t] _lowerCamelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) _lowerCamelCase = 1 - alpha_prod_t _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] _lowerCamelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output _lowerCamelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) _lowerCamelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def UpperCamelCase_ ( A_ , A_ , A_ ) -> torch.Tensor: """simple docstring""" _lowerCamelCase = acos(torch.dot(torch.flatten(A_ ) , torch.flatten(A_ ) ) / torch.norm(A_ ) / torch.norm(A_ ) ) return sin((1 - alpha) * theta ) * xa / sin(A_ ) + sin(alpha * theta ) * xa / sin(A_ )
638
1
snake_case__ = {} def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on _lowerCamelCase = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one _lowerCamelCase = _calculate(days - 1 , __UpperCAmelCase , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 _lowerCamelCase = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter _lowerCamelCase = _calculate(days - 1 , __UpperCAmelCase , 0 ) _lowerCamelCase = state_late + state_absent + state_ontime _lowerCamelCase = prizestrings return prizestrings def __magic_name__( __UpperCAmelCase = 30 ) -> int: '''simple docstring''' return _calculate(__UpperCAmelCase , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
638
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowercase ) , 'Tatoeba directory does not exist.' ) class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=A_ ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=A_ ) assert mmeta["long_pair"] == "heb-eng"
638
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = { 'google/mobilenet_v2_1.4_224': 'https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json', 'google/mobilenet_v2_1.0_224': 'https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json', 'google/mobilenet_v2_0.75_160': 'https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json', 'google/mobilenet_v2_0.35_96': 'https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'mobilenet_v2' def __init__( self , A_=3 , A_=2_24 , A_=1.0 , A_=8 , A_=8 , A_=6 , A_=32 , A_=True , A_=True , A_="relu6" , A_=True , A_=0.8 , A_=0.02 , A_=0.001 , A_=2_55 , **A_ , ) -> str: """simple docstring""" super().__init__(**A_ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) _lowerCamelCase = num_channels _lowerCamelCase = image_size _lowerCamelCase = depth_multiplier _lowerCamelCase = depth_divisible_by _lowerCamelCase = min_depth _lowerCamelCase = expand_ratio _lowerCamelCase = output_stride _lowerCamelCase = first_layer_is_expansion _lowerCamelCase = finegrained_output _lowerCamelCase = hidden_act _lowerCamelCase = tf_padding _lowerCamelCase = classifier_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = semantic_loss_ignore_index class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = version.parse('1.11' ) @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def UpperCamelCase_ ( self ) -> float: """simple docstring""" return 1E-4
638
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' _lowerCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase = in_proj_bias[: config.hidden_size] _lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase = in_proj_bias[-config.hidden_size :] def __magic_name__( __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True ) -> str: '''simple docstring''' _lowerCamelCase = ViTConfig() # patch_size if model_name[-1] == "8": _lowerCamelCase = 8 # set labels if required if not base_model: _lowerCamelCase = 1000 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _lowerCamelCase = 384 _lowerCamelCase = 1536 _lowerCamelCase = 12 _lowerCamelCase = 6 # load original model from torch hub _lowerCamelCase = torch.hub.load('''facebookresearch/dino:main''' , __UpperCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase = original_model.state_dict() if base_model: remove_classification_head_(__UpperCAmelCase ) _lowerCamelCase = create_rename_keys(__UpperCAmelCase , base_model=__UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # load HuggingFace model if base_model: _lowerCamelCase = ViTModel(__UpperCAmelCase , add_pooling_layer=__UpperCAmelCase ).eval() else: _lowerCamelCase = ViTForImageClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _lowerCamelCase = ViTImageProcessor() _lowerCamelCase = image_processor(images=prepare_img() , return_tensors='''pt''' ) _lowerCamelCase = encoding['''pixel_values'''] _lowerCamelCase = model(__UpperCAmelCase ) if base_model: _lowerCamelCase = original_model(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _lowerCamelCase = original_model(__UpperCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) snake_case__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
638
1
from __future__ import annotations from math import ceil, floor, sqrt def __magic_name__( __UpperCAmelCase = 200_0000 ) -> int: '''simple docstring''' _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(__UpperCAmelCase ) _lowerCamelCase = ceil(__UpperCAmelCase ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f'''{solution() = }''')
638
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _lowerCamelCase = '''lm_head''' _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): _lowerCamelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True ) -> Union[str, Any]: '''simple docstring''' if config_path is not None: _lowerCamelCase = UniSpeechConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _lowerCamelCase = Dictionary.load_from_json(__UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _lowerCamelCase = target_dict.pad_index _lowerCamelCase = target_dict.bos_index _lowerCamelCase = target_dict.eos_index _lowerCamelCase = len(target_dict.symbols ) _lowerCamelCase = os.path.join(__UpperCAmelCase , '''vocab.json''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__UpperCAmelCase ) ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) _lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _lowerCamelCase = 42 _lowerCamelCase = 43 with open(__UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = WavaVecaPhonemeCTCTokenizer( __UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__UpperCAmelCase , ) _lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False _lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) _lowerCamelCase = WavaVecaProcessor(feature_extractor=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) _lowerCamelCase = UniSpeechForCTC(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechForPreTraining(__UpperCAmelCase ) if is_finetuned: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _lowerCamelCase = model[0].eval() recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) hf_unispeech.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
638
1
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) snake_case__ = logging.getLogger(__name__) snake_case__ = tf.data.AUTOTUNE def __magic_name__( ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser(description='''Train a masked language model on TPU.''' ) parser.add_argument( '''--pretrained_model_config''' , type=__UpperCAmelCase , default='''roberta-base''' , help='''The model config to use. Note that we don\'t copy the model\'s weights, only the config!''' , ) parser.add_argument( '''--tokenizer''' , type=__UpperCAmelCase , default='''unigram-tokenizer-wikitext''' , help='''The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model\'s vocab size.''' , ) parser.add_argument( '''--per_replica_batch_size''' , type=__UpperCAmelCase , default=8 , help='''Batch size per TPU core.''' , ) parser.add_argument( '''--no_tpu''' , action='''store_true''' , help='''If set, run on CPU and don\'t try to initialize a TPU. Useful for debugging on non-TPU instances.''' , ) parser.add_argument( '''--tpu_name''' , type=__UpperCAmelCase , help='''Name of TPU resource to initialize. Should be blank on Colab, and \'local\' on TPU VMs.''' , default='''local''' , ) parser.add_argument( '''--tpu_zone''' , type=__UpperCAmelCase , help='''Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes.''' , ) parser.add_argument( '''--gcp_project''' , type=__UpperCAmelCase , help='''Google cloud project name. Only used for non-Colab TPU nodes.''' ) parser.add_argument( '''--bfloat16''' , action='''store_true''' , help='''Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU.''' , ) parser.add_argument( '''--train_dataset''' , type=__UpperCAmelCase , help='''Path to training dataset to load. If the path begins with `gs://`''' ''' then the dataset will be loaded from a Google Cloud Storage bucket.''' , ) parser.add_argument( '''--shuffle_buffer_size''' , type=__UpperCAmelCase , default=2**18 , help='''Size of the shuffle buffer (in samples)''' , ) parser.add_argument( '''--eval_dataset''' , type=__UpperCAmelCase , help='''Path to evaluation dataset to load. If the path begins with `gs://`''' ''' then the dataset will be loaded from a Google Cloud Storage bucket.''' , ) parser.add_argument( '''--num_epochs''' , type=__UpperCAmelCase , default=1 , help='''Number of epochs to train for.''' , ) parser.add_argument( '''--learning_rate''' , type=__UpperCAmelCase , default=1E-4 , help='''Learning rate to use for training.''' , ) parser.add_argument( '''--weight_decay_rate''' , type=__UpperCAmelCase , default=1E-3 , help='''Weight decay rate to use for training.''' , ) parser.add_argument( '''--max_length''' , type=__UpperCAmelCase , default=512 , help='''Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py''' , ) parser.add_argument( '''--mlm_probability''' , type=__UpperCAmelCase , default=0.1_5 , help='''Fraction of tokens to mask during training.''' , ) parser.add_argument('''--output_dir''' , type=__UpperCAmelCase , required=__UpperCAmelCase , help='''Path to save model checkpoints to.''' ) parser.add_argument('''--hub_model_id''' , type=__UpperCAmelCase , help='''Model ID to upload to on the Hugging Face Hub.''' ) _lowerCamelCase = parser.parse_args() return args def __magic_name__( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' try: if args.tpu_name: _lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: _lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( '''Couldn\'t connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or ''' '''--gcp_project. When running on a TPU VM, use --tpu_name local.''' ) tf.config.experimental_connect_to_cluster(__UpperCAmelCase ) tf.tpu.experimental.initialize_tpu_system(__UpperCAmelCase ) return tpu def __magic_name__( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase = 0 for file in file_list: _lowerCamelCase = file.split('''/''' )[-1] _lowerCamelCase = re.search(r'''-\d+-(\d+)\.tfrecord''' , __UpperCAmelCase ).group(1 ) _lowerCamelCase = int(__UpperCAmelCase ) num_samples += sample_count return num_samples def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ) -> str: '''simple docstring''' _lowerCamelCase = count_samples(__UpperCAmelCase ) _lowerCamelCase = tf.data.Dataset.from_tensor_slices(__UpperCAmelCase ) if shuffle: _lowerCamelCase = dataset.shuffle(len(__UpperCAmelCase ) ) _lowerCamelCase = tf.data.TFRecordDataset(__UpperCAmelCase , num_parallel_reads=__UpperCAmelCase ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here _lowerCamelCase = dataset.apply(tf.data.experimental.assert_cardinality(__UpperCAmelCase ) ) _lowerCamelCase = dataset.map(__UpperCAmelCase , num_parallel_calls=__UpperCAmelCase ) if shuffle: assert shuffle_buffer_size is not None _lowerCamelCase = dataset.shuffle(args.shuffle_buffer_size ) _lowerCamelCase = dataset.batch(__UpperCAmelCase , drop_remainder=__UpperCAmelCase ) _lowerCamelCase = dataset.map(__UpperCAmelCase , num_parallel_calls=__UpperCAmelCase ) _lowerCamelCase = dataset.prefetch(__UpperCAmelCase ) return dataset def __magic_name__( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if not args.no_tpu: _lowerCamelCase = initialize_tpu(__UpperCAmelCase ) _lowerCamelCase = tf.distribute.TPUStrategy(__UpperCAmelCase ) else: _lowerCamelCase = tf.distribute.OneDeviceStrategy(device='''/gpu:0''' ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy('''mixed_bfloat16''' ) _lowerCamelCase = AutoTokenizer.from_pretrained(args.tokenizer ) _lowerCamelCase = AutoConfig.from_pretrained(args.pretrained_model_config ) _lowerCamelCase = tokenizer.vocab_size _lowerCamelCase = tf.io.gfile.glob(os.path.join(args.train_dataset , '''*.tfrecord''' ) ) if not training_records: raise ValueError(F'No .tfrecord files found in {args.train_dataset}.' ) _lowerCamelCase = tf.io.gfile.glob(os.path.join(args.eval_dataset , '''*.tfrecord''' ) ) if not eval_records: raise ValueError(F'No .tfrecord files found in {args.eval_dataset}.' ) _lowerCamelCase = count_samples(__UpperCAmelCase ) _lowerCamelCase = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) _lowerCamelCase = steps_per_epoch * args.num_epochs with strategy.scope(): _lowerCamelCase = TFAutoModelForMaskedLM.from_config(__UpperCAmelCase ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built _lowerCamelCase , _lowerCamelCase = create_optimizer( num_train_steps=__UpperCAmelCase , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=__UpperCAmelCase , metrics=['''accuracy'''] ) def decode_fn(__UpperCAmelCase ): _lowerCamelCase = { '''input_ids''': tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), '''attention_mask''': tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(__UpperCAmelCase , __UpperCAmelCase ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. _lowerCamelCase = DataCollatorForLanguageModeling( tokenizer=__UpperCAmelCase , mlm_probability=args.mlm_probability , mlm=__UpperCAmelCase , return_tensors='''tf''' ) def mask_with_collator(__UpperCAmelCase ): # TF really needs an isin() function _lowerCamelCase = ( ~tf.cast(batch['''attention_mask'''] , tf.bool ) | (batch['''input_ids'''] == tokenizer.cls_token_id) | (batch['''input_ids'''] == tokenizer.sep_token_id) ) _lowerCamelCase , _lowerCamelCase = data_collator.tf_mask_tokens( batch['''input_ids'''] , vocab_size=len(__UpperCAmelCase ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=__UpperCAmelCase , ) return batch _lowerCamelCase = args.per_replica_batch_size * strategy.num_replicas_in_sync _lowerCamelCase = prepare_dataset( __UpperCAmelCase , decode_fn=__UpperCAmelCase , mask_fn=__UpperCAmelCase , batch_size=__UpperCAmelCase , shuffle=__UpperCAmelCase , shuffle_buffer_size=args.shuffle_buffer_size , ) _lowerCamelCase = prepare_dataset( __UpperCAmelCase , decode_fn=__UpperCAmelCase , mask_fn=__UpperCAmelCase , batch_size=__UpperCAmelCase , shuffle=__UpperCAmelCase , ) _lowerCamelCase = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=__UpperCAmelCase ) ) model.fit( __UpperCAmelCase , validation_data=__UpperCAmelCase , epochs=args.num_epochs , callbacks=__UpperCAmelCase , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": snake_case__ = parse_args() main(args)
638
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> None: """simple docstring""" warnings.warn( '''The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DPTImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
638
1
import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaTokenizerFast, ) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = model.config _lowerCamelCase = DonutSwinConfig( image_size=original_config.input_size , patch_size=4 , depths=original_config.encoder_layer , num_heads=[4, 8, 16, 32] , window_size=original_config.window_size , embed_dim=128 , ) _lowerCamelCase = MBartConfig( is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , add_cross_attention=__UpperCAmelCase , decoder_layers=original_config.decoder_layer , max_position_embeddings=original_config.max_position_embeddings , vocab_size=len( model.decoder.tokenizer ) , scale_embedding=__UpperCAmelCase , add_final_layer_norm=__UpperCAmelCase , ) return encoder_config, decoder_config def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' if "encoder.model" in name: _lowerCamelCase = name.replace('''encoder.model''' , '''encoder''' ) if "decoder.model" in name: _lowerCamelCase = name.replace('''decoder.model''' , '''decoder''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if name.startswith('''encoder''' ): if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "mask" not in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": _lowerCamelCase = '''encoder.layernorm.weight''' if name == "encoder.norm.bias": _lowerCamelCase = '''encoder.layernorm.bias''' return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = int(key_split[5] ) _lowerCamelCase = model.encoder.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] elif "attn_mask" in key or key in ["encoder.model.norm.weight", "encoder.model.norm.bias"]: # HuggingFace implementation doesn't use attn_mask buffer # and model doesn't use final LayerNorms for the encoder pass else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = DonutModel.from_pretrained(__UpperCAmelCase ).eval() # load HuggingFace model _lowerCamelCase , _lowerCamelCase = get_configs(__UpperCAmelCase ) _lowerCamelCase = DonutSwinModel(__UpperCAmelCase ) _lowerCamelCase = MBartForCausalLM(__UpperCAmelCase ) _lowerCamelCase = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() _lowerCamelCase = original_model.state_dict() _lowerCamelCase = convert_state_dict(__UpperCAmelCase , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) # verify results on scanned document _lowerCamelCase = load_dataset('''hf-internal-testing/example-documents''' ) _lowerCamelCase = dataset['''test'''][0]['''image'''].convert('''RGB''' ) _lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained(__UpperCAmelCase , from_slow=__UpperCAmelCase ) _lowerCamelCase = DonutImageProcessor( do_align_long_axis=original_model.config.align_long_axis , size=original_model.config.input_size[::-1] ) _lowerCamelCase = DonutProcessor(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values if model_name == "naver-clova-ix/donut-base-finetuned-docvqa": _lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' _lowerCamelCase = '''When is the coffee break?''' _lowerCamelCase = task_prompt.replace('''{user_input}''' , __UpperCAmelCase ) elif model_name == "naver-clova-ix/donut-base-finetuned-rvlcdip": _lowerCamelCase = '''<s_rvlcdip>''' elif model_name in [ "naver-clova-ix/donut-base-finetuned-cord-v1", "naver-clova-ix/donut-base-finetuned-cord-v1-2560", ]: _lowerCamelCase = '''<s_cord>''' elif model_name == "naver-clova-ix/donut-base-finetuned-cord-v2": _lowerCamelCase = '''s_cord-v2>''' elif model_name == "naver-clova-ix/donut-base-finetuned-zhtrainticket": _lowerCamelCase = '''<s_zhtrainticket>''' elif model_name in ["naver-clova-ix/donut-proto", "naver-clova-ix/donut-base"]: # use a random prompt _lowerCamelCase = '''hello world''' else: raise ValueError('''Model name not supported''' ) _lowerCamelCase = original_model.decoder.tokenizer(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors='''pt''' )[ '''input_ids''' ] _lowerCamelCase = original_model.encoder.model.patch_embed(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = model.encoder.embeddings(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) # verify encoder hidden states _lowerCamelCase = original_model.encoder(__UpperCAmelCase ) _lowerCamelCase = model.encoder(__UpperCAmelCase ).last_hidden_state assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) # verify decoder hidden states _lowerCamelCase = original_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ).logits _lowerCamelCase = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: model.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) processor.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='naver-clova-ix/donut-base-finetuned-docvqa', required=False, type=str, help='Name of the original model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, required=False, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub.', ) snake_case__ = parser.parse_args() convert_donut_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
import argparse import json import subprocess def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = ( F'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) _lowerCamelCase = subprocess.run(__UpperCAmelCase , shell=__UpperCAmelCase , stdout=subprocess.PIPE ) _lowerCamelCase = output.stdout.decode('''utf-8''' ) _lowerCamelCase = json.loads(__UpperCAmelCase ) _lowerCamelCase = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__UpperCAmelCase ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) ) if len(__UpperCAmelCase ) > 0: _lowerCamelCase = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' return values.split(''',''' ) snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) snake_case__ = parser.parse_args() get_runner_status(args.target_runners, args.token)
638
1
from __future__ import annotations snake_case__ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class UpperCamelCase : '''simple docstring''' def __init__( self , A_ , A_ ) -> None: """simple docstring""" _lowerCamelCase = graph # mapping node to its parent in resulting breadth first tree _lowerCamelCase = {} _lowerCamelCase = source_vertex def UpperCamelCase_ ( self ) -> None: """simple docstring""" _lowerCamelCase = {self.source_vertex} _lowerCamelCase = None _lowerCamelCase = [self.source_vertex] # first in first out queue while queue: _lowerCamelCase = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(A_ ) _lowerCamelCase = vertex queue.append(A_ ) def UpperCamelCase_ ( self , A_ ) -> str: """simple docstring""" if target_vertex == self.source_vertex: return self.source_vertex _lowerCamelCase = self.parent.get(A_ ) if target_vertex_parent is None: _lowerCamelCase = ( F'No path from vertex: {self.source_vertex} to vertex: {target_vertex}' ) raise ValueError(A_ ) return self.shortest_path(A_ ) + F'->{target_vertex}' if __name__ == "__main__": snake_case__ = Graph(graph, 'G') g.breath_first_search() print(g.shortest_path('D')) print(g.shortest_path('G')) print(g.shortest_path('Foo'))
638
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
638
1
from typing import Tuple, Union from ...modeling_outputs import BackboneOutput from ...modeling_utils import PreTrainedModel from ...utils import is_timm_available, is_torch_available, requires_backends from ...utils.backbone_utils import BackboneMixin from .configuration_timm_backbone import TimmBackboneConfig if is_timm_available(): import timm if is_torch_available(): from torch import Tensor class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' A_ = 'pixel_values' A_ = False A_ = TimmBackboneConfig def __init__( self , A_ , **A_ ) -> Any: """simple docstring""" requires_backends(self , '''timm''' ) super().__init__(A_ ) _lowerCamelCase = config if config.backbone is None: raise ValueError('''backbone is not set in the config. Please set it to a timm model name.''' ) if config.backbone not in timm.list_models(): raise ValueError(F'backbone {config.backbone} is not supported by timm.' ) if hasattr(A_ , '''out_features''' ) and config.out_features is not None: raise ValueError('''out_features is not supported by TimmBackbone. Please use out_indices instead.''' ) _lowerCamelCase = getattr(A_ , '''use_pretrained_backbone''' , A_ ) if pretrained is None: raise ValueError('''use_pretrained_backbone is not set in the config. Please set it to True or False.''' ) # We just take the final layer by default. This matches the default for the transformers models. _lowerCamelCase = config.out_indices if getattr(A_ , '''out_indices''' , A_ ) is not None else (-1,) _lowerCamelCase = timm.create_model( config.backbone , pretrained=A_ , features_only=config.features_only , in_chans=config.num_channels , out_indices=A_ , **A_ , ) # These are used to control the output of the model when called. If output_hidden_states is True, then # return_layers is modified to include all layers. _lowerCamelCase = self._backbone.return_layers _lowerCamelCase = {layer['''module''']: str(A_ ) for i, layer in enumerate(self._backbone.feature_info.info )} super()._init_backbone(A_ ) @classmethod def UpperCamelCase_ ( cls , A_ , *A_ , **A_ ) -> Dict: """simple docstring""" requires_backends(cls , ['''vision''', '''timm'''] ) from ...models.timm_backbone import TimmBackboneConfig _lowerCamelCase = kwargs.pop('''config''' , TimmBackboneConfig() ) _lowerCamelCase = kwargs.pop('''use_timm_backbone''' , A_ ) if not use_timm: raise ValueError('''use_timm_backbone must be True for timm backbones''' ) _lowerCamelCase = kwargs.pop('''num_channels''' , config.num_channels ) _lowerCamelCase = kwargs.pop('''features_only''' , config.features_only ) _lowerCamelCase = kwargs.pop('''use_pretrained_backbone''' , config.use_pretrained_backbone ) _lowerCamelCase = kwargs.pop('''out_indices''' , config.out_indices ) _lowerCamelCase = TimmBackboneConfig( backbone=A_ , num_channels=A_ , features_only=A_ , use_pretrained_backbone=A_ , out_indices=A_ , ) return super()._from_config(A_ , **A_ ) def UpperCamelCase_ ( self , A_ ) -> Any: """simple docstring""" pass def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=None , **A_ ) -> Union[BackboneOutput, Tuple[Tensor, ...]]: """simple docstring""" _lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict _lowerCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCamelCase = output_attentions if output_attentions is not None else self.config.output_attentions if output_attentions: raise ValueError('''Cannot output attentions for timm backbones at the moment''' ) if output_hidden_states: # We modify the return layers to include all the stages of the backbone _lowerCamelCase = self._all_layers _lowerCamelCase = self._backbone(A_ , **A_ ) _lowerCamelCase = self._return_layers _lowerCamelCase = tuple(hidden_states[i] for i in self.out_indices ) else: _lowerCamelCase = self._backbone(A_ , **A_ ) _lowerCamelCase = None _lowerCamelCase = tuple(A_ ) _lowerCamelCase = tuple(A_ ) if hidden_states is not None else None if not return_dict: _lowerCamelCase = (feature_maps,) if output_hidden_states: _lowerCamelCase = output + (hidden_states,) return output return BackboneOutput(feature_maps=A_ , hidden_states=A_ , attentions=A_ )
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(__UpperCAmelCase , n - 1 , __UpperCAmelCase ) * a) % mod else: _lowerCamelCase = binary_exponentiation(__UpperCAmelCase , n / 2 , __UpperCAmelCase ) return (b * b) % mod # a prime number snake_case__ = 701 snake_case__ = 10_0000_0000 snake_case__ = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
638
import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaTokenizerFast, ) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = model.config _lowerCamelCase = DonutSwinConfig( image_size=original_config.input_size , patch_size=4 , depths=original_config.encoder_layer , num_heads=[4, 8, 16, 32] , window_size=original_config.window_size , embed_dim=128 , ) _lowerCamelCase = MBartConfig( is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , add_cross_attention=__UpperCAmelCase , decoder_layers=original_config.decoder_layer , max_position_embeddings=original_config.max_position_embeddings , vocab_size=len( model.decoder.tokenizer ) , scale_embedding=__UpperCAmelCase , add_final_layer_norm=__UpperCAmelCase , ) return encoder_config, decoder_config def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' if "encoder.model" in name: _lowerCamelCase = name.replace('''encoder.model''' , '''encoder''' ) if "decoder.model" in name: _lowerCamelCase = name.replace('''decoder.model''' , '''decoder''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if name.startswith('''encoder''' ): if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "mask" not in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": _lowerCamelCase = '''encoder.layernorm.weight''' if name == "encoder.norm.bias": _lowerCamelCase = '''encoder.layernorm.bias''' return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = int(key_split[5] ) _lowerCamelCase = model.encoder.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] elif "attn_mask" in key or key in ["encoder.model.norm.weight", "encoder.model.norm.bias"]: # HuggingFace implementation doesn't use attn_mask buffer # and model doesn't use final LayerNorms for the encoder pass else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = DonutModel.from_pretrained(__UpperCAmelCase ).eval() # load HuggingFace model _lowerCamelCase , _lowerCamelCase = get_configs(__UpperCAmelCase ) _lowerCamelCase = DonutSwinModel(__UpperCAmelCase ) _lowerCamelCase = MBartForCausalLM(__UpperCAmelCase ) _lowerCamelCase = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() _lowerCamelCase = original_model.state_dict() _lowerCamelCase = convert_state_dict(__UpperCAmelCase , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) # verify results on scanned document _lowerCamelCase = load_dataset('''hf-internal-testing/example-documents''' ) _lowerCamelCase = dataset['''test'''][0]['''image'''].convert('''RGB''' ) _lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained(__UpperCAmelCase , from_slow=__UpperCAmelCase ) _lowerCamelCase = DonutImageProcessor( do_align_long_axis=original_model.config.align_long_axis , size=original_model.config.input_size[::-1] ) _lowerCamelCase = DonutProcessor(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values if model_name == "naver-clova-ix/donut-base-finetuned-docvqa": _lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' _lowerCamelCase = '''When is the coffee break?''' _lowerCamelCase = task_prompt.replace('''{user_input}''' , __UpperCAmelCase ) elif model_name == "naver-clova-ix/donut-base-finetuned-rvlcdip": _lowerCamelCase = '''<s_rvlcdip>''' elif model_name in [ "naver-clova-ix/donut-base-finetuned-cord-v1", "naver-clova-ix/donut-base-finetuned-cord-v1-2560", ]: _lowerCamelCase = '''<s_cord>''' elif model_name == "naver-clova-ix/donut-base-finetuned-cord-v2": _lowerCamelCase = '''s_cord-v2>''' elif model_name == "naver-clova-ix/donut-base-finetuned-zhtrainticket": _lowerCamelCase = '''<s_zhtrainticket>''' elif model_name in ["naver-clova-ix/donut-proto", "naver-clova-ix/donut-base"]: # use a random prompt _lowerCamelCase = '''hello world''' else: raise ValueError('''Model name not supported''' ) _lowerCamelCase = original_model.decoder.tokenizer(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors='''pt''' )[ '''input_ids''' ] _lowerCamelCase = original_model.encoder.model.patch_embed(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = model.encoder.embeddings(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) # verify encoder hidden states _lowerCamelCase = original_model.encoder(__UpperCAmelCase ) _lowerCamelCase = model.encoder(__UpperCAmelCase ).last_hidden_state assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) # verify decoder hidden states _lowerCamelCase = original_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ).logits _lowerCamelCase = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: model.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) processor.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='naver-clova-ix/donut-base-finetuned-docvqa', required=False, type=str, help='Name of the original model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, required=False, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub.', ) snake_case__ = parser.parse_args() convert_donut_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
1
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
638
1
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors snake_case__ = logging.getLogger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'sequence-classification' def __init__( self , A_ ) -> str: """simple docstring""" if type(A_ ) == dict: _lowerCamelCase = Namespace(**A_ ) _lowerCamelCase = glue_output_modes[hparams.task] _lowerCamelCase = glue_tasks_num_labels[hparams.task] super().__init__(A_ , A_ , self.mode ) def UpperCamelCase_ ( self , **A_ ) -> Optional[int]: """simple docstring""" return self.model(**A_ ) def UpperCamelCase_ ( self , A_ , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: _lowerCamelCase = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None _lowerCamelCase = self(**A_ ) _lowerCamelCase = outputs[0] _lowerCamelCase = self.trainer.lr_schedulers[0]['''scheduler'''] _lowerCamelCase = {'''loss''': loss, '''rate''': lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.hparams _lowerCamelCase = processors[args.task]() _lowerCamelCase = processor.get_labels() for mode in ["train", "dev"]: _lowerCamelCase = self._feature_file(A_ ) if os.path.exists(A_ ) and not args.overwrite_cache: logger.info('''Loading features from cached file %s''' , A_ ) else: logger.info('''Creating features from dataset file at %s''' , args.data_dir ) _lowerCamelCase = ( processor.get_dev_examples(args.data_dir ) if mode == '''dev''' else processor.get_train_examples(args.data_dir ) ) _lowerCamelCase = convert_examples_to_features( A_ , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info('''Saving features into cached file %s''' , A_ ) torch.save(A_ , A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_ = False ) -> DataLoader: """simple docstring""" _lowerCamelCase = '''dev''' if mode == '''test''' else mode _lowerCamelCase = self._feature_file(A_ ) logger.info('''Loading features from cached file %s''' , A_ ) _lowerCamelCase = torch.load(A_ ) _lowerCamelCase = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) _lowerCamelCase = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) _lowerCamelCase = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": _lowerCamelCase = torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": _lowerCamelCase = torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(A_ , A_ , A_ , A_ ) , batch_size=A_ , shuffle=A_ , ) def UpperCamelCase_ ( self , A_ , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: _lowerCamelCase = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None _lowerCamelCase = self(**A_ ) _lowerCamelCase , _lowerCamelCase = outputs[:2] _lowerCamelCase = logits.detach().cpu().numpy() _lowerCamelCase = inputs['''labels'''].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase_ ( self , A_ ) -> tuple: """simple docstring""" _lowerCamelCase = torch.stack([x['''val_loss'''] for x in outputs] ).mean().detach().cpu().item() _lowerCamelCase = np.concatenate([x['''pred'''] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": _lowerCamelCase = np.argmax(A_ , axis=1 ) elif self.hparams.glue_output_mode == "regression": _lowerCamelCase = np.squeeze(A_ ) _lowerCamelCase = np.concatenate([x['''target'''] for x in outputs] , axis=0 ) _lowerCamelCase = [[] for _ in range(out_label_ids.shape[0] )] _lowerCamelCase = [[] for _ in range(out_label_ids.shape[0] )] _lowerCamelCase = {**{'''val_loss''': val_loss_mean}, **compute_metrics(self.hparams.task , A_ , A_ )} _lowerCamelCase = dict(results.items() ) _lowerCamelCase = results return ret, preds_list, out_label_list def UpperCamelCase_ ( self , A_ ) -> dict: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self._eval_end(A_ ) _lowerCamelCase = ret['''log'''] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase_ ( self , A_ ) -> dict: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self._eval_end(A_ ) _lowerCamelCase = ret['''log'''] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" BaseTransformer.add_model_specific_args(A_ , A_ ) parser.add_argument( '''--max_seq_length''' , default=1_28 , type=A_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--task''' , default='''''' , type=A_ , required=A_ , help='''The GLUE task to run''' , ) parser.add_argument( '''--gpus''' , default=0 , type=A_ , help='''The number of GPUs allocated for this, it is by default 0 meaning none''' , ) parser.add_argument( '''--overwrite_cache''' , action='''store_true''' , help='''Overwrite the cached training and evaluation sets''' ) return parser def __magic_name__( ) -> Dict: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() add_generic_args(__UpperCAmelCase , os.getcwd() ) _lowerCamelCase = GLUETransformer.add_model_specific_args(__UpperCAmelCase , os.getcwd() ) _lowerCamelCase = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: _lowerCamelCase = os.path.join( '''./results''' , F'{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}' , ) os.makedirs(args.output_dir ) _lowerCamelCase = GLUETransformer(__UpperCAmelCase ) _lowerCamelCase = generic_train(__UpperCAmelCase , __UpperCAmelCase ) # Optionally, predict on dev set and write to output_dir if args.do_predict: _lowerCamelCase = sorted(glob.glob(os.path.join(args.output_dir , '''checkpoint-epoch=*.ckpt''' ) , recursive=__UpperCAmelCase ) ) _lowerCamelCase = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(__UpperCAmelCase ) if __name__ == "__main__": main()
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
from __future__ import annotations import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' return np.maximum(0 , __UpperCAmelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
638
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
1
import collections import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = '▁' snake_case__ = {'vocab_file': 'prophetnet.tokenizer'} snake_case__ = { 'vocab_file': { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/prophetnet.tokenizer' ), } } snake_case__ = { 'microsoft/xprophetnet-large-wiki100-cased': {'do_lower_case': False}, } snake_case__ = { 'microsoft/xprophetnet-large-wiki100-cased': 512, } def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' _lowerCamelCase = collections.OrderedDict() with open(__UpperCAmelCase , '''r''' , encoding='''utf-8''' ) as reader: _lowerCamelCase = reader.readlines() for index, token in enumerate(__UpperCAmelCase ): _lowerCamelCase = token.rstrip('''\n''' ) _lowerCamelCase = index return vocab class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['input_ids', 'attention_mask'] def __init__( self , A_ , A_="[SEP]" , A_="[SEP]" , A_="[SEP]" , A_="[UNK]" , A_="[PAD]" , A_="[CLS]" , A_="[MASK]" , A_ = None , **A_ , ) -> None: """simple docstring""" _lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A_ , eos_token=A_ , sep_token=A_ , unk_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) try: import sentencepiece as spm except ImportError: logger.warning( '''You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece''' ''' pip install sentencepiece''' ) raise _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(A_ ) ) _lowerCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # put special tokens and [unused] tokens into the vocab _lowerCamelCase = {'''[PAD]''': 0, '''[CLS]''': 1, '''[SEP]''': 2, '''[UNK]''': 3, '''[MASK]''': 4} for i in range(10 ): _lowerCamelCase = F'[unused{i}]' _lowerCamelCase = 5 + i # The first "real" token "," has position 15 in the embedding vocab and position 3 in the spm vocab _lowerCamelCase = 12 _lowerCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} for k in self.fairseq_tokens_to_ids.keys(): self.unique_no_split_tokens.append(A_ ) def __getstate__( self ) -> List[str]: """simple docstring""" _lowerCamelCase = self.__dict__.copy() _lowerCamelCase = None return state def __setstate__( self , A_ ) -> Any: """simple docstring""" _lowerCamelCase = d try: import sentencepiece as spm except ImportError: logger.warning( '''You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece''' ''' pip install sentencepiece''' ) raise # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _lowerCamelCase = {} _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase_ ( self , A_ , A_ = None , A_ = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is None: return ([0] * len(A_ )) + [1] return ([0] * len(A_ )) + [1] + ([0] * len(A_ )) + [1] def UpperCamelCase_ ( self , A_ , A_ = None ) -> List[int]: """simple docstring""" _lowerCamelCase = [self.sep_token_id] if token_ids_a is None: return len(token_ids_a + sep ) * [0] return len(token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return len(self.sp_model ) + self.fairseq_offset def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase_ ( self , A_ ) -> str: """simple docstring""" return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase_ ( self , A_ ) -> Optional[Any]: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCamelCase = self.sp_model.PieceToId(A_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def UpperCamelCase_ ( self , A_ ) -> Tuple: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def UpperCamelCase_ ( self , A_ ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = ''''''.join(A_ ).replace(A_ , ''' ''' ).strip() return out_string def UpperCamelCase_ ( self , A_ , A_ = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(A_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCamelCase = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: _lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase_ ( self , A_ , A_ = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.sep_token_id] _lowerCamelCase = [self.sep_token_id] return token_ids_a + sep + token_ids_a + sep
638
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
1
import warnings from functools import wraps from typing import Callable def __magic_name__( __UpperCAmelCase ) -> Callable: '''simple docstring''' @wraps(__UpperCAmelCase ) def _inner_fn(*__UpperCAmelCase , **__UpperCAmelCase ): warnings.warn( (F'\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.') , __UpperCAmelCase , ) return fn(*__UpperCAmelCase , **__UpperCAmelCase ) return _inner_fn
638
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
1
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand snake_case__ = logging.get_logger(__name__) # pylint: disable=invalid-name def __magic_name__( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__UpperCAmelCase ): return ext raise Exception( F'Unable to determine file format from file extension {path}. ' F'Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}' ) def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _lowerCamelCase = try_infer_format_from_ext(args.input ) if args.format == '''infer''' else args.format _lowerCamelCase = PipelineDataFormat.from_str( format=__UpperCAmelCase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__UpperCAmelCase , __UpperCAmelCase ) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , A_ , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = nlp _lowerCamelCase = reader @staticmethod def UpperCamelCase_ ( A_ ) -> Dict: """simple docstring""" _lowerCamelCase = parser.add_parser('''run''' , help='''Run a pipeline through the CLI''' ) run_parser.add_argument('''--task''' , choices=get_supported_tasks() , help='''Task to run''' ) run_parser.add_argument('''--input''' , type=A_ , help='''Path to the file to use for inference''' ) run_parser.add_argument('''--output''' , type=A_ , help='''Path to the file that will be used post to write results.''' ) run_parser.add_argument('''--model''' , type=A_ , help='''Name or path to the model to instantiate.''' ) run_parser.add_argument('''--config''' , type=A_ , help='''Name or path to the model\'s config to instantiate.''' ) run_parser.add_argument( '''--tokenizer''' , type=A_ , help='''Name of the tokenizer to use. (default: same as the model name)''' ) run_parser.add_argument( '''--column''' , type=A_ , help='''Name of the column to use as input. (For multi columns input as QA use column1,columns2)''' , ) run_parser.add_argument( '''--format''' , type=A_ , default='''infer''' , choices=PipelineDataFormat.SUPPORTED_FORMATS , help='''Input format to read from''' , ) run_parser.add_argument( '''--device''' , type=A_ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) run_parser.add_argument('''--overwrite''' , action='''store_true''' , help='''Allow overwriting the output file.''' ) run_parser.set_defaults(func=A_ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self._nlp, [] for entry in self._reader: _lowerCamelCase = nlp(**A_ ) if self._reader.is_multi_columns else nlp(A_ ) if isinstance(A_ , A_ ): outputs.append(A_ ) else: outputs += output # Saving data if self._nlp.binary_output: _lowerCamelCase = self._reader.save_binary(A_ ) logger.warning(F'Current pipeline requires output to be in binary format, saving at {binary_path}' ) else: self._reader.save(A_ )
638
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } _lowerCamelCase = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub _lowerCamelCase = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase_ ( self , **A_ ) -> str: """simple docstring""" _lowerCamelCase = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> Optional[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> int: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = feature_extractor(A_ , return_tensors='''np''' ) _lowerCamelCase = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = '''This is a test string''' _lowerCamelCase = processor(text=A_ ) _lowerCamelCase = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase_ ( self , A_=(2, 10, 16) , A_=77 ) -> Optional[Any]: """simple docstring""" np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits(shape=(10, 16) , seed=13 ) _lowerCamelCase = processor.decode(A_ ) _lowerCamelCase = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: _lowerCamelCase = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: _lowerCamelCase = processor.batch_decode(A_ , A_ ) _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as p: _lowerCamelCase = decoder.decode_beams_batch(A_ , A_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 15 _lowerCamelCase = -20.0 _lowerCamelCase = -4.0 _lowerCamelCase = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] _lowerCamelCase = [d[0][2] for d in decoded_decoder_out] _lowerCamelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 2.0 _lowerCamelCase = 5.0 _lowerCamelCase = -20.0 _lowerCamelCase = True _lowerCamelCase = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = snapshot_download('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = processor_wavaveca(A_ , return_tensors='''np''' ) _lowerCamelCase = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor_wavaveca.batch_decode(A_ ) _lowerCamelCase = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = [d[key] for d in offsets] return retrieved_list def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits()[0] _lowerCamelCase = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" import torch _lowerCamelCase = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) _lowerCamelCase = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) _lowerCamelCase = iter(A_ ) _lowerCamelCase = next(A_ ) _lowerCamelCase = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) _lowerCamelCase = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train _lowerCamelCase = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): _lowerCamelCase = model(A_ ).logits.cpu().numpy() _lowerCamelCase = processor.decode(logits[0] , output_word_offsets=A_ ) _lowerCamelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate _lowerCamelCase = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] _lowerCamelCase = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off _lowerCamelCase = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) _lowerCamelCase = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
638
1
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class UpperCamelCase : '''simple docstring''' def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=64 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=5_12 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ) -> Tuple: """simple docstring""" _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_input_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope _lowerCamelCase = vocab_size - 1 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase = True return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = GPTNeoXModel(config=A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ ) _lowerCamelCase = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = True _lowerCamelCase = GPTNeoXModel(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = GPTNeoXForCausalLM(config=A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = self.num_labels _lowerCamelCase = GPTNeoXForQuestionAnswering(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self.num_labels _lowerCamelCase = GPTNeoXForSequenceClassification(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.num_labels _lowerCamelCase = GPTNeoXForTokenClassification(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = True _lowerCamelCase = GPTNeoXForCausalLM(config=A_ ) model.to(A_ ) model.eval() # first forward pass _lowerCamelCase = model(A_ , attention_mask=A_ , use_cache=A_ ) _lowerCamelCase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCamelCase = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = torch.cat([input_mask, next_mask] , dim=-1 ) _lowerCamelCase = model(A_ , attention_mask=A_ , output_hidden_states=A_ ) _lowerCamelCase = output_from_no_past['''hidden_states'''][0] _lowerCamelCase = model( A_ , attention_mask=A_ , past_key_values=A_ , output_hidden_states=A_ , )['''hidden_states'''][0] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A_ , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( __lowercase , __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) A_ = (GPTNeoXForCausalLM,) if is_torch_available() else () A_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) A_ = False A_ = False A_ = False A_ = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = GPTNeoXModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=A_ , hidden_size=64 , num_attention_heads=8 ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # This regression test was failing with PyTorch < 1.3 _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() _lowerCamelCase = None self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*A_ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @unittest.skip(reason='''Feed forward chunking is not implemented''' ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass @parameterized.expand([('''linear''',), ('''dynamic''',)] ) def UpperCamelCase_ ( self , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ids_tensor([1, 10] , config.vocab_size ) _lowerCamelCase = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCamelCase = GPTNeoXModel(A_ ) original_model.to(A_ ) original_model.eval() _lowerCamelCase = original_model(A_ ).last_hidden_state _lowerCamelCase = original_model(A_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCamelCase = {'''type''': scaling_type, '''factor''': 10.0} _lowerCamelCase = GPTNeoXModel(A_ ) scaled_model.to(A_ ) scaled_model.eval() _lowerCamelCase = scaled_model(A_ ).last_hidden_state _lowerCamelCase = scaled_model(A_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(A_ , A_ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(A_ , A_ , atol=1E-5 ) ) @require_torch class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = AutoTokenizer.from_pretrained('''EleutherAI/pythia-410m-deduped''' ) for checkpointing in [True, False]: _lowerCamelCase = GPTNeoXForCausalLM.from_pretrained('''EleutherAI/pythia-410m-deduped''' ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(A_ ) _lowerCamelCase = tokenizer('''My favorite food is''' , return_tensors='''pt''' ).to(A_ ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 _lowerCamelCase = '''My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI\'m not sure''' _lowerCamelCase = model.generate(**A_ , do_sample=A_ , max_new_tokens=20 ) _lowerCamelCase = tokenizer.batch_decode(A_ )[0] self.assertEqual(A_ , A_ )
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _lowerCamelCase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _lowerCamelCase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _lowerCamelCase = subset[i - 1][j] if arr[i - 1] <= j: _lowerCamelCase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' if len(__UpperCAmelCase ) != len(__UpperCAmelCase ): raise ValueError('''The length of profit and weight must be same.''' ) if max_weight <= 0: raise ValueError('''max_weight must greater than zero.''' ) if any(p < 0 for p in profit ): raise ValueError('''Profit can not be negative.''' ) if any(w < 0 for w in weight ): raise ValueError('''Weight can not be negative.''' ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. _lowerCamelCase = [p / w for p, w in zip(__UpperCAmelCase , __UpperCAmelCase )] # Creating a copy of the list and sorting profit/weight in ascending order _lowerCamelCase = sorted(__UpperCAmelCase ) # declaring useful variables _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight _lowerCamelCase = sorted_profit_by_weight[length - i - 1] _lowerCamelCase = profit_by_weight.index(__UpperCAmelCase ) _lowerCamelCase = -1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( 'Input profits, weights, and then max_weight (all positive ints) separated by ' 'spaces.' ) snake_case__ = [int(x) for x in input('Input profits separated by spaces: ').split()] snake_case__ = [int(x) for x in input('Input weights separated by spaces: ').split()] snake_case__ = int(input('Max weight allowed: ')) # Function Call calc_profit(profit, weight, max_weight)
638
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _lowerCamelCase = '''lm_head''' _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): _lowerCamelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True ) -> Union[str, Any]: '''simple docstring''' if config_path is not None: _lowerCamelCase = UniSpeechConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _lowerCamelCase = Dictionary.load_from_json(__UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _lowerCamelCase = target_dict.pad_index _lowerCamelCase = target_dict.bos_index _lowerCamelCase = target_dict.eos_index _lowerCamelCase = len(target_dict.symbols ) _lowerCamelCase = os.path.join(__UpperCAmelCase , '''vocab.json''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__UpperCAmelCase ) ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) _lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _lowerCamelCase = 42 _lowerCamelCase = 43 with open(__UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = WavaVecaPhonemeCTCTokenizer( __UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__UpperCAmelCase , ) _lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False _lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) _lowerCamelCase = WavaVecaProcessor(feature_extractor=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) _lowerCamelCase = UniSpeechForCTC(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechForPreTraining(__UpperCAmelCase ) if is_finetuned: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _lowerCamelCase = model[0].eval() recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) hf_unispeech.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
638
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
1
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ) -> None: """simple docstring""" _lowerCamelCase = device _lowerCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _lowerCamelCase = [0.48145466, 0.4578275, 0.40821073] _lowerCamelCase = [0.26862954, 0.26130258, 0.27577711] _lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCamelCase = torchvision.transforms.Resize(2_24 ) _lowerCamelCase = torchvision.transforms.CenterCrop(2_24 ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.resize(A_ ) _lowerCamelCase = self.center_crop(A_ ) _lowerCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer(text=A_ , **A_ ) _lowerCamelCase = self.preprocess_img(A_ ) _lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ) -> None: """simple docstring""" super().__init__() _lowerCamelCase = None _lowerCamelCase = device if device else get_device() if vqgan: _lowerCamelCase = vqgan else: _lowerCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _lowerCamelCase = clip else: _lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _lowerCamelCase = ProcessorGradientFlow(device=self.device ) _lowerCamelCase = iterations _lowerCamelCase = lr _lowerCamelCase = log _lowerCamelCase = make_grid _lowerCamelCase = return_val _lowerCamelCase = quantize _lowerCamelCase = self.vqgan.decoder.z_shape def UpperCamelCase_ ( self , A_=None , A_=None , A_=5 , A_=True ) -> Any: """simple docstring""" _lowerCamelCase = [] if output_path is None: _lowerCamelCase = '''./animation.gif''' if input_path is None: _lowerCamelCase = self.save_path _lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(A_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(A_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _lowerCamelCase = total_duration / len(A_ ) _lowerCamelCase = [frame_duration] * len(A_ ) if extend_frames: _lowerCamelCase = 1.5 _lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F'gif saved to {output_path}' ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> Union[str, Any]: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _lowerCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _lowerCamelCase = preprocess_vqgan(A_ ) _lowerCamelCase , *_lowerCamelCase = self.vqgan.encode(A_ ) return z def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.latent.detach().requires_grad_() _lowerCamelCase = base_latent + transform_vector if self.quantize: _lowerCamelCase , *_lowerCamelCase = self.vqgan.quantize(A_ ) else: _lowerCamelCase = trans_latent return self.vqgan.decode(A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_=None ) -> Any: """simple docstring""" _lowerCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors='''pt''' , padding=A_ ) _lowerCamelCase = self.clip(**A_ ) _lowerCamelCase = clip_outputs.logits_per_image if weights is not None: _lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , A_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , A_ , weights=neg_prompts['''weights'''] ) else: _lowerCamelCase = torch.tensor([1] , device=self.device ) _lowerCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCamelCase = self._add_vector(A_ ) _lowerCamelCase = loop_post_process(A_ ) _lowerCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print('''CLIP loss''' , A_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" wandb.init(reinit=A_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _lowerCamelCase = Image.open(A_ ) _lowerCamelCase = image.resize((2_56, 2_56) ) wandb.log('''Original Image''' , wandb.Image(A_ ) ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if not prompts: return [] _lowerCamelCase = [] _lowerCamelCase = [] if isinstance(A_ , A_ ): _lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _lowerCamelCase = prompt[0] _lowerCamelCase = float(prompt[1] ) elif ":" in prompt: _lowerCamelCase , _lowerCamelCase = prompt.split(''':''' ) _lowerCamelCase = float(A_ ) else: _lowerCamelCase = prompt _lowerCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ) -> str: """simple docstring""" if image_path: _lowerCamelCase = self._get_latent(A_ ) else: _lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCamelCase = self.process_prompts(A_ ) _lowerCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(A_ ) _lowerCamelCase = save_path _lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(A_ ) ) _lowerCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({'''Image''': wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
638
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available snake_case__ = { 'configuration_poolformer': [ 'POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PoolFormerConfig', 'PoolFormerOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['PoolFormerFeatureExtractor'] snake_case__ = ['PoolFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PoolFormerForImageClassification', 'PoolFormerModel', 'PoolFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure)
638
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case__ = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='%(message)s') def __magic_name__( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size, 1) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.mean(1 ) # Centralize the data of class i _lowerCamelCase = data - column_reshape(__UpperCAmelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(__UpperCAmelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = features.mean(1 ) _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.shape[1] _lowerCamelCase = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' if features.any(): _lowerCamelCase = features.mean(1 ) # Center the dataset _lowerCamelCase = features - np.reshape(__UpperCAmelCase , (data_mean.size, 1) ) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) / features.shape[1] _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(__UpperCAmelCase ) # Take all the columns in the reverse order (-1), and then takes only the first _lowerCamelCase = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space _lowerCamelCase = np.dot(filtered_eigenvectors.T , __UpperCAmelCase ) logging.info('''Principal Component Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' assert classes > dimensions # Check if features have been already loaded if features.any: _lowerCamelCase , _lowerCamelCase = eigh( covariance_between_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , covariance_within_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , ) _lowerCamelCase = eigenvectors[:, ::-1][:, :dimensions] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = np.linalg.svd(__UpperCAmelCase ) _lowerCamelCase = svd_matrix[:, 0:dimensions] _lowerCamelCase = np.dot(filtered_svd_matrix.T , __UpperCAmelCase ) logging.info('''Linear Discriminant Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) _lowerCamelCase = np.array([0, 0, 0, 1, 1] ) _lowerCamelCase = 2 _lowerCamelCase = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = linear_discriminant_analysis( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if isinstance(__UpperCAmelCase , np.ndarray ): raise AssertionError( '''Did not raise AssertionError for dimensions > classes''' ) assert error_info.type is AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) _lowerCamelCase = 2 _lowerCamelCase = np.array([[6.9_2_8_2_0_3_2_3, 8.6_6_0_2_5_4_0_4, 1_0.3_9_2_3_0_4_8_5], [3.0, 3.0, 3.0]] ) with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = principal_component_analysis(__UpperCAmelCase , __UpperCAmelCase ) if not np.allclose(__UpperCAmelCase , __UpperCAmelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
638
1
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow snake_case__ = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) snake_case__ = logging.getLogger() def __magic_name__( ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''-f''' ) _lowerCamelCase = parser.parse_args() return args.f def __magic_name__( __UpperCAmelCase , __UpperCAmelCase="eval" ) -> List[str]: '''simple docstring''' _lowerCamelCase = os.path.join(__UpperCAmelCase , F'{split}_results.json' ) if os.path.exists(__UpperCAmelCase ): with open(__UpperCAmelCase , '''r''' ) as f: return json.load(__UpperCAmelCase ) raise ValueError(F'can\'t find {path}' ) snake_case__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class UpperCamelCase ( __lowercase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_glue.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --eval_steps=2\n --warmup_steps=2\n --seed=42\n --max_seq_length=128\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_flax_glue.main() _lowerCamelCase = get_results(A_ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) @slow def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_clm_flax.py\n --model_name_or_path distilgpt2\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --do_train\n --do_eval\n --block_size 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --num_train_epochs 2\n --logging_steps 2 --eval_steps 2\n --output_dir {tmp_dir}\n --overwrite_output_dir\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_clm_flax.main() _lowerCamelCase = get_results(A_ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_summarization.py\n --model_name_or_path t5-small\n --train_file tests/fixtures/tests_samples/xsum/sample.json\n --validation_file tests/fixtures/tests_samples/xsum/sample.json\n --test_file tests/fixtures/tests_samples/xsum/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --num_train_epochs=3\n --warmup_steps=8\n --do_train\n --do_eval\n --do_predict\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --predict_with_generate\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_summarization_flax.main() _lowerCamelCase = get_results(A_ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_mlm.py\n --model_name_or_path distilroberta-base\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --max_seq_length 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --logging_steps 2 --eval_steps 2\n --do_train\n --do_eval\n --num_train_epochs=1\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_mlm_flax.main() _lowerCamelCase = get_results(A_ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_t5_mlm_flax.py\n --model_name_or_path t5-small\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --do_train\n --do_eval\n --max_seq_length 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --num_train_epochs 2\n --logging_steps 2 --eval_steps 2\n --output_dir {tmp_dir}\n --overwrite_output_dir\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_ta_mlm_flax.main() _lowerCamelCase = get_results(A_ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.42 ) @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu _lowerCamelCase = 7 if get_gpu_count() > 1 else 2 _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_flax_ner.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/conll/sample.json\n --validation_file tests/fixtures/tests_samples/conll/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --do_train\n --do_eval\n --warmup_steps=2\n --learning_rate=2e-4\n --logging_steps 2 --eval_steps 2\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=2\n --num_train_epochs={epochs}\n --seed 7\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_flax_ner.main() _lowerCamelCase = get_results(A_ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_auto_remove_tmp_dir() _lowerCamelCase = F'\n run_qa.py\n --model_name_or_path bert-base-uncased\n --version_2_with_negative\n --train_file tests/fixtures/tests_samples/SQUAD/sample.json\n --validation_file tests/fixtures/tests_samples/SQUAD/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --num_train_epochs=3\n --warmup_steps=2\n --do_train\n --do_eval\n --logging_steps 2 --eval_steps 2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n '.split() with patch.object(A_ , '''argv''' , A_ ): run_qa.main() _lowerCamelCase = get_results(A_ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
638
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['vqvae'] def __init__( self , A_ , A_ , A_ , A_ , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules(unet=A_ , scheduler=A_ , mel=A_ , vqvae=A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 50 if isinstance(self.scheduler , A_ ) else 10_00 @torch.no_grad() def __call__( self , A_ = 1 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = 0 , A_ = None , A_ = None , A_=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: """simple docstring""" _lowerCamelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(A_ ) _lowerCamelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: _lowerCamelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: _lowerCamelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=A_ , device=self.device , ) _lowerCamelCase = noise _lowerCamelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A_ , A_ ) _lowerCamelCase = self.mel.audio_slice_to_image(A_ ) _lowerCamelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) _lowerCamelCase = (input_image / 2_55) * 2 - 1 _lowerCamelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: _lowerCamelCase = self.vqvae.encode(torch.unsqueeze(A_ , 0 ) ).latent_dist.sample( generator=A_ )[0] _lowerCamelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , self.scheduler.timesteps[start_step - 1] ) _lowerCamelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) _lowerCamelCase = int(mask_start_secs * pixels_per_second ) _lowerCamelCase = int(mask_end_secs * pixels_per_second ) _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , A_ ): _lowerCamelCase = self.unet(A_ , A_ , A_ )['''sample'''] else: _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] if isinstance(self.scheduler , A_ ): _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , eta=A_ , generator=A_ , )['''prev_sample'''] else: _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , generator=A_ , )['''prev_sample'''] if mask is not None: if mask_start > 0: _lowerCamelCase = mask[:, step, :, :mask_start] if mask_end > 0: _lowerCamelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance _lowerCamelCase = 1 / self.vqvae.config.scaling_factor * images _lowerCamelCase = self.vqvae.decode(A_ )['''sample'''] _lowerCamelCase = (images / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() _lowerCamelCase = (images * 2_55).round().astype('''uint8''' ) _lowerCamelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A_ , mode='''RGB''' ).convert('''L''' ) for _ in images) ) _lowerCamelCase = [self.mel.image_to_audio(A_ ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A_ )[:, np.newaxis, :] ) , **ImagePipelineOutput(A_ ) ) @torch.no_grad() def UpperCamelCase_ ( self , A_ , A_ = 50 ) -> np.ndarray: """simple docstring""" assert isinstance(self.scheduler , A_ ) self.scheduler.set_timesteps(A_ ) _lowerCamelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) _lowerCamelCase = (sample / 2_55) * 2 - 1 _lowerCamelCase = torch.Tensor(A_ ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): _lowerCamelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps _lowerCamelCase = self.scheduler.alphas_cumprod[t] _lowerCamelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) _lowerCamelCase = 1 - alpha_prod_t _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] _lowerCamelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output _lowerCamelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) _lowerCamelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def UpperCamelCase_ ( A_ , A_ , A_ ) -> torch.Tensor: """simple docstring""" _lowerCamelCase = acos(torch.dot(torch.flatten(A_ ) , torch.flatten(A_ ) ) / torch.norm(A_ ) / torch.norm(A_ ) ) return sin((1 - alpha) * theta ) * xa / sin(A_ ) + sin(alpha * theta ) * xa / sin(A_ )
638
1
from __future__ import annotations snake_case__ = 8.988E9 # units = N * m^s * C^-2 def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> dict[str, float]: '''simple docstring''' _lowerCamelCase = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if distance < 0: raise ValueError('''Distance cannot be negative''' ) if force == 0: _lowerCamelCase = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: _lowerCamelCase = abs(__UpperCAmelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: _lowerCamelCase = abs(__UpperCAmelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: _lowerCamelCase = (COULOMBS_CONSTANT * charge_product / abs(__UpperCAmelCase )) ** 0.5 return {"distance": distance} raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
638
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowercase ) , 'Tatoeba directory does not exist.' ) class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=A_ ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=A_ ) assert mmeta["long_pair"] == "heb-eng"
638
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = TFCamembertModel.from_pretrained('''jplu/tf-camembert-base''' ) _lowerCamelCase = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _lowerCamelCase = model(A_ )['''last_hidden_state'''] _lowerCamelCase = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , A_ ) # compare the actual values for a slice. _lowerCamelCase = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
638
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' _lowerCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase = in_proj_bias[: config.hidden_size] _lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase = in_proj_bias[-config.hidden_size :] def __magic_name__( __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True ) -> str: '''simple docstring''' _lowerCamelCase = ViTConfig() # patch_size if model_name[-1] == "8": _lowerCamelCase = 8 # set labels if required if not base_model: _lowerCamelCase = 1000 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _lowerCamelCase = 384 _lowerCamelCase = 1536 _lowerCamelCase = 12 _lowerCamelCase = 6 # load original model from torch hub _lowerCamelCase = torch.hub.load('''facebookresearch/dino:main''' , __UpperCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase = original_model.state_dict() if base_model: remove_classification_head_(__UpperCAmelCase ) _lowerCamelCase = create_rename_keys(__UpperCAmelCase , base_model=__UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # load HuggingFace model if base_model: _lowerCamelCase = ViTModel(__UpperCAmelCase , add_pooling_layer=__UpperCAmelCase ).eval() else: _lowerCamelCase = ViTForImageClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _lowerCamelCase = ViTImageProcessor() _lowerCamelCase = image_processor(images=prepare_img() , return_tensors='''pt''' ) _lowerCamelCase = encoding['''pixel_values'''] _lowerCamelCase = model(__UpperCAmelCase ) if base_model: _lowerCamelCase = original_model(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _lowerCamelCase = original_model(__UpperCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) snake_case__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
638
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = [] # fmt: off # stem: rename_keys.append(('''cls_token''', '''vit.embeddings.cls_token''') ) rename_keys.append(('''pos_embed''', '''vit.embeddings.position_embeddings''') ) rename_keys.append(('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias''') ) # backbone rename_keys.append(('''patch_embed.backbone.stem.conv.weight''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight''') ) rename_keys.append(('''patch_embed.backbone.stem.norm.weight''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight''') ) rename_keys.append(('''patch_embed.backbone.stem.norm.bias''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias''') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) # fmt: on return rename_keys def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> Tuple: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase = in_proj_bias[: config.hidden_size] _lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase = in_proj_bias[-config.hidden_size :] def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' _lowerCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = BitConfig( global_padding='''same''' , layer_type='''bottleneck''' , depths=(3, 4, 9) , out_features=['''stage3'''] , embedding_dynamic_padding=__UpperCAmelCase , ) _lowerCamelCase = ViTHybridConfig(backbone_config=__UpperCAmelCase , image_size=384 , num_labels=1000 ) _lowerCamelCase = False # load original model from timm _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__UpperCAmelCase ) _lowerCamelCase = create_rename_keys(__UpperCAmelCase , __UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": _lowerCamelCase = ViTHybridModel(__UpperCAmelCase ).eval() else: _lowerCamelCase = ViTHybridForImageClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # create image processor _lowerCamelCase = create_transform(**resolve_data_config({} , model=__UpperCAmelCase ) ) _lowerCamelCase = transform.transforms _lowerCamelCase = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } _lowerCamelCase = ViTHybridImageProcessor( do_resize=__UpperCAmelCase , size={'''shortest_edge''': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=__UpperCAmelCase , crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} , do_normalize=__UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) _lowerCamelCase = prepare_img() _lowerCamelCase = transform(__UpperCAmelCase ).unsqueeze(0 ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase ) # verify logits with torch.no_grad(): _lowerCamelCase = model(__UpperCAmelCase ) _lowerCamelCase = outputs.logits print('''Predicted class:''' , logits.argmax(-1 ).item() ) if base_model: _lowerCamelCase = timm_model.forward_features(__UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: _lowerCamelCase = timm_model(__UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F'Saving model {vit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: print(F'Pushing model and processor to the hub {vit_name}' ) model.push_to_hub(F'ybelkada/{vit_name}' ) processor.push_to_hub(F'ybelkada/{vit_name}' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) snake_case__ = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _lowerCamelCase = '''lm_head''' _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): _lowerCamelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True ) -> Union[str, Any]: '''simple docstring''' if config_path is not None: _lowerCamelCase = UniSpeechConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _lowerCamelCase = Dictionary.load_from_json(__UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _lowerCamelCase = target_dict.pad_index _lowerCamelCase = target_dict.bos_index _lowerCamelCase = target_dict.eos_index _lowerCamelCase = len(target_dict.symbols ) _lowerCamelCase = os.path.join(__UpperCAmelCase , '''vocab.json''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__UpperCAmelCase ) ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) _lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _lowerCamelCase = 42 _lowerCamelCase = 43 with open(__UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = WavaVecaPhonemeCTCTokenizer( __UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__UpperCAmelCase , ) _lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False _lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) _lowerCamelCase = WavaVecaProcessor(feature_extractor=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) _lowerCamelCase = UniSpeechForCTC(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechForPreTraining(__UpperCAmelCase ) if is_finetuned: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _lowerCamelCase = model[0].eval() recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) hf_unispeech.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
638
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowercase ) , 'Tatoeba directory does not exist.' ) class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=A_ ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=A_ ) assert mmeta["long_pair"] == "heb-eng"
638
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> None: """simple docstring""" warnings.warn( '''The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DPTImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
638
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class UpperCamelCase ( __lowercase ): '''simple docstring''' @staticmethod @abstractmethod def UpperCamelCase_ ( A_ ) -> Union[str, Any]: """simple docstring""" raise NotImplementedError() @abstractmethod def UpperCamelCase_ ( self ) -> str: """simple docstring""" raise NotImplementedError()
638
import argparse import json import subprocess def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = ( F'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) _lowerCamelCase = subprocess.run(__UpperCAmelCase , shell=__UpperCAmelCase , stdout=subprocess.PIPE ) _lowerCamelCase = output.stdout.decode('''utf-8''' ) _lowerCamelCase = json.loads(__UpperCAmelCase ) _lowerCamelCase = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__UpperCAmelCase ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) ) if len(__UpperCAmelCase ) > 0: _lowerCamelCase = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' return values.split(''',''' ) snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) snake_case__ = parser.parse_args() get_runner_status(args.target_runners, args.token)
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' return int((input_a, input_a).count(1 ) != 0 ) def __magic_name__( ) -> None: '''simple docstring''' assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
638
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
638
1
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
1
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' monkeypatch.setattr('''datasets.utils.deprecation_utils._emitted_deprecation_warnings''' , set() ) @pytest.fixture def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' class UpperCamelCase : '''simple docstring''' def __init__( self , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = metric_id class UpperCamelCase : '''simple docstring''' A_ = [MetricMock(__lowercase ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self._metrics monkeypatch.setattr('''datasets.inspect.huggingface_hub''' , HfhMock() ) @pytest.mark.parametrize( '''func, args''' , [(load_metric, ('''metrics/mse''',)), (list_metrics, ()), (inspect_metric, ('''metrics/mse''', '''tmp_path'''))] ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' if "tmp_path" in args: _lowerCamelCase = tuple(arg if arg != '''tmp_path''' else tmp_path for arg in args ) with pytest.warns(__UpperCAmelCase , match='''https://huggingface.co/docs/evaluate''' ): func(*__UpperCAmelCase )
638
import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaTokenizerFast, ) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = model.config _lowerCamelCase = DonutSwinConfig( image_size=original_config.input_size , patch_size=4 , depths=original_config.encoder_layer , num_heads=[4, 8, 16, 32] , window_size=original_config.window_size , embed_dim=128 , ) _lowerCamelCase = MBartConfig( is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , add_cross_attention=__UpperCAmelCase , decoder_layers=original_config.decoder_layer , max_position_embeddings=original_config.max_position_embeddings , vocab_size=len( model.decoder.tokenizer ) , scale_embedding=__UpperCAmelCase , add_final_layer_norm=__UpperCAmelCase , ) return encoder_config, decoder_config def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' if "encoder.model" in name: _lowerCamelCase = name.replace('''encoder.model''' , '''encoder''' ) if "decoder.model" in name: _lowerCamelCase = name.replace('''decoder.model''' , '''decoder''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if name.startswith('''encoder''' ): if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "mask" not in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": _lowerCamelCase = '''encoder.layernorm.weight''' if name == "encoder.norm.bias": _lowerCamelCase = '''encoder.layernorm.bias''' return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = int(key_split[5] ) _lowerCamelCase = model.encoder.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] elif "attn_mask" in key or key in ["encoder.model.norm.weight", "encoder.model.norm.bias"]: # HuggingFace implementation doesn't use attn_mask buffer # and model doesn't use final LayerNorms for the encoder pass else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = DonutModel.from_pretrained(__UpperCAmelCase ).eval() # load HuggingFace model _lowerCamelCase , _lowerCamelCase = get_configs(__UpperCAmelCase ) _lowerCamelCase = DonutSwinModel(__UpperCAmelCase ) _lowerCamelCase = MBartForCausalLM(__UpperCAmelCase ) _lowerCamelCase = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() _lowerCamelCase = original_model.state_dict() _lowerCamelCase = convert_state_dict(__UpperCAmelCase , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) # verify results on scanned document _lowerCamelCase = load_dataset('''hf-internal-testing/example-documents''' ) _lowerCamelCase = dataset['''test'''][0]['''image'''].convert('''RGB''' ) _lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained(__UpperCAmelCase , from_slow=__UpperCAmelCase ) _lowerCamelCase = DonutImageProcessor( do_align_long_axis=original_model.config.align_long_axis , size=original_model.config.input_size[::-1] ) _lowerCamelCase = DonutProcessor(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values if model_name == "naver-clova-ix/donut-base-finetuned-docvqa": _lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' _lowerCamelCase = '''When is the coffee break?''' _lowerCamelCase = task_prompt.replace('''{user_input}''' , __UpperCAmelCase ) elif model_name == "naver-clova-ix/donut-base-finetuned-rvlcdip": _lowerCamelCase = '''<s_rvlcdip>''' elif model_name in [ "naver-clova-ix/donut-base-finetuned-cord-v1", "naver-clova-ix/donut-base-finetuned-cord-v1-2560", ]: _lowerCamelCase = '''<s_cord>''' elif model_name == "naver-clova-ix/donut-base-finetuned-cord-v2": _lowerCamelCase = '''s_cord-v2>''' elif model_name == "naver-clova-ix/donut-base-finetuned-zhtrainticket": _lowerCamelCase = '''<s_zhtrainticket>''' elif model_name in ["naver-clova-ix/donut-proto", "naver-clova-ix/donut-base"]: # use a random prompt _lowerCamelCase = '''hello world''' else: raise ValueError('''Model name not supported''' ) _lowerCamelCase = original_model.decoder.tokenizer(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors='''pt''' )[ '''input_ids''' ] _lowerCamelCase = original_model.encoder.model.patch_embed(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = model.encoder.embeddings(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) # verify encoder hidden states _lowerCamelCase = original_model.encoder(__UpperCAmelCase ) _lowerCamelCase = model.encoder(__UpperCAmelCase ).last_hidden_state assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) # verify decoder hidden states _lowerCamelCase = original_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ).logits _lowerCamelCase = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: model.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) processor.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='naver-clova-ix/donut-base-finetuned-docvqa', required=False, type=str, help='Name of the original model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, required=False, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub.', ) snake_case__ = parser.parse_args() convert_donut_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
1
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING snake_case__ = logging.get_logger(__name__) snake_case__ = Dict[str, Any] snake_case__ = List[Prediction] @add_end_docstrings(__lowercase ) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , '''vision''' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def UpperCamelCase_ ( self , **A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = {} if "threshold" in kwargs: _lowerCamelCase = kwargs['''threshold'''] return {}, {}, postprocess_kwargs def __call__( self , *A_ , **A_ ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def UpperCamelCase_ ( self , A_ ) -> Tuple: """simple docstring""" _lowerCamelCase = load_image(A_ ) _lowerCamelCase = torch.IntTensor([[image.height, image.width]] ) _lowerCamelCase = self.image_processor(images=[image] , return_tensors='''pt''' ) if self.tokenizer is not None: _lowerCamelCase = self.tokenizer(text=inputs['''words'''] , boxes=inputs['''boxes'''] , return_tensors='''pt''' ) _lowerCamelCase = target_size return inputs def UpperCamelCase_ ( self , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = model_inputs.pop('''target_size''' ) _lowerCamelCase = self.model(**A_ ) _lowerCamelCase = outputs.__class__({'''target_size''': target_size, **outputs} ) if self.tokenizer is not None: _lowerCamelCase = model_inputs['''bbox'''] return model_outputs def UpperCamelCase_ ( self , A_ , A_=0.9 ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = model_outputs['''target_size'''] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. _lowerCamelCase , _lowerCamelCase = target_size[0].tolist() def unnormalize(A_ ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 10_00), (height * bbox[1] / 10_00), (width * bbox[2] / 10_00), (height * bbox[3] / 10_00), ] ) ) _lowerCamelCase , _lowerCamelCase = model_outputs['''logits'''].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) _lowerCamelCase = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] _lowerCamelCase = [unnormalize(A_ ) for bbox in model_outputs['''bbox'''].squeeze(0 )] _lowerCamelCase = ['''score''', '''label''', '''box'''] _lowerCamelCase = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel _lowerCamelCase = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) _lowerCamelCase = raw_annotations[0] _lowerCamelCase = raw_annotation['''scores'''] _lowerCamelCase = raw_annotation['''labels'''] _lowerCamelCase = raw_annotation['''boxes'''] _lowerCamelCase = scores.tolist() _lowerCamelCase = [self.model.config.idalabel[label.item()] for label in labels] _lowerCamelCase = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] _lowerCamelCase = ['''score''', '''label''', '''box'''] _lowerCamelCase = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['''scores'''] , raw_annotation['''labels'''] , raw_annotation['''boxes'''] ) ] return annotation def UpperCamelCase_ ( self , A_ ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('''The ObjectDetectionPipeline is only available in PyTorch.''' ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = box.int().tolist() _lowerCamelCase = { '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
638
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
638
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device snake_case__ = False class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' pass @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=A_ , generator=A_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A_ ) _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = generator.manual_seed(0 ) _lowerCamelCase = pipe( prompt=A_ , generator=A_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=A_ , generator=A_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images _lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowerCamelCase = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self , A_ ) -> List[Any]: """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): _lowerCamelCase = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(A_ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=A_ , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = '''sgugger/tiny-distilbert-classification''' _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , only_pretrain_model=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = AutoConfig.from_pretrained(A_ ) _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=A_ , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ , [config] ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = AutoConfig.from_pretrained(A_ ) _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ , [config] ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = AutoConfig.from_pretrained(A_ ) _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ , [config] ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = '''patrickvonplaten/t5-tiny-random''' _lowerCamelCase = AutoConfig.from_pretrained(A_ ) _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ , configs=[config] ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , '''Cannot do xla on CPU.''' ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=A_ , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , use_xla=A_ , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=A_ , save_to_csv=A_ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(A_ , '''inf_time.csv''' ) , inference_memory_csv_file=os.path.join(A_ , '''inf_mem.csv''' ) , env_info_csv_file=os.path.join(A_ , '''env.csv''' ) , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) benchmark.run() self.assertTrue(Path(os.path.join(A_ , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ , '''env.csv''' ) ).exists() ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(A_ ): self.assertTrue(hasattr(A_ , '''sequential''' ) ) self.assertTrue(hasattr(A_ , '''cumulative''' ) ) self.assertTrue(hasattr(A_ , '''current''' ) ) self.assertTrue(hasattr(A_ , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=A_ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(A_ , '''log.txt''' ) , log_print=A_ , trace_memory_line_by_line=A_ , eager_mode=A_ , multi_process=A_ , ) _lowerCamelCase = TensorFlowBenchmark(A_ ) _lowerCamelCase = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(A_ , '''log.txt''' ) ).exists() )
638
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
1
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values snake_case__ = argparse.ArgumentParser() parser.add_argument('--user', type=str, default='ubuntu') parser.add_argument('--host', type=str, default='localhost') parser.add_argument('--key_path', type=str, default=None) parser.add_argument('--instance', type=str, default='V100:1') parser.add_argument('--provider', type=str, default='cheapest') parser.add_argument('--use_spot', type=bool, default=False) parser.add_argument('--example', type=str, default='pytorch/text-generation/run_generation.py') snake_case__ , snake_case__ = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('Cannot specify both BYO and on-demand cluster args') snake_case__ = rh.cluster( name='rh-cluster', ips=[args.host], ssh_creds={'ssh_user': args.user, 'ssh_private_key': args.key_path} ) else: snake_case__ = rh.cluster( name='rh-cluster', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) snake_case__ = args.example.rsplit('/', 1)[0] # Set up remote environment cluster.install_packages(['pip:./']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f'''pip install -r transformers/examples/{example_dir}/requirements.txt''']) cluster.run(['pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f'''python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}''']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
638
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for attribute in key.split('''.''' ): _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name and "relative_attention_bias" not in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = torch.load(__UpperCAmelCase ) _lowerCamelCase = WavLMConfigOrig(checkpoint['''cfg'''] ) _lowerCamelCase = WavLMOrig(__UpperCAmelCase ) model.load_state_dict(checkpoint['''model'''] ) model.eval() if config_path is not None: _lowerCamelCase = WavLMConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = WavLMConfig() _lowerCamelCase = WavLMModel(__UpperCAmelCase ) recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase ) hf_wavlm.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') snake_case__ = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
638
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING snake_case__ = logging.get_logger(__name__) snake_case__ = { 'salesforce/blip2-opt-2.7b': 'https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json', } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'blip_2_vision_model' def __init__( self , A_=14_08 , A_=61_44 , A_=39 , A_=16 , A_=2_24 , A_=14 , A_="gelu" , A_=0.00001 , A_=0.0 , A_=1E-1_0 , A_=True , **A_ , ) -> int: """simple docstring""" super().__init__(**A_ ) _lowerCamelCase = hidden_size _lowerCamelCase = intermediate_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = patch_size _lowerCamelCase = image_size _lowerCamelCase = initializer_range _lowerCamelCase = attention_dropout _lowerCamelCase = layer_norm_eps _lowerCamelCase = hidden_act _lowerCamelCase = qkv_bias @classmethod def UpperCamelCase_ ( cls , A_ , **A_ ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(A_ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(A_ , **A_ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get('''model_type''' ) == "blip-2": _lowerCamelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A_ , **A_ ) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'blip_2_qformer' def __init__( self , A_=3_05_22 , A_=7_68 , A_=12 , A_=12 , A_=30_72 , A_="gelu" , A_=0.1 , A_=0.1 , A_=5_12 , A_=0.02 , A_=1E-1_2 , A_=0 , A_="absolute" , A_=2 , A_=14_08 , **A_ , ) -> Union[str, Any]: """simple docstring""" super().__init__(pad_token_id=A_ , **A_ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = position_embedding_type _lowerCamelCase = cross_attention_frequency _lowerCamelCase = encoder_hidden_size @classmethod def UpperCamelCase_ ( cls , A_ , **A_ ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(A_ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(A_ , **A_ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get('''model_type''' ) == "blip-2": _lowerCamelCase = config_dict['''qformer_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A_ , **A_ ) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'blip-2' A_ = True def __init__( self , A_=None , A_=None , A_=None , A_=32 , **A_ ) -> str: """simple docstring""" super().__init__(**A_ ) if vision_config is None: _lowerCamelCase = {} logger.info('''vision_config is None. initializing the Blip2VisionConfig with default values.''' ) if qformer_config is None: _lowerCamelCase = {} logger.info('''qformer_config is None. Initializing the Blip2QFormerConfig with default values.''' ) if text_config is None: _lowerCamelCase = {} logger.info('''text_config is None. Initializing the text config with default values (`OPTConfig`).''' ) _lowerCamelCase = BlipaVisionConfig(**A_ ) _lowerCamelCase = BlipaQFormerConfig(**A_ ) _lowerCamelCase = text_config['''model_type'''] if '''model_type''' in text_config else '''opt''' _lowerCamelCase = CONFIG_MAPPING[text_model_type](**A_ ) _lowerCamelCase = self.text_config.tie_word_embeddings _lowerCamelCase = self.text_config.is_encoder_decoder _lowerCamelCase = num_query_tokens _lowerCamelCase = self.vision_config.hidden_size _lowerCamelCase = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES _lowerCamelCase = 1.0 _lowerCamelCase = 0.02 @classmethod def UpperCamelCase_ ( cls , A_ , A_ , A_ , **A_ , ) -> Tuple: """simple docstring""" return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **A_ , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.vision_config.to_dict() _lowerCamelCase = self.qformer_config.to_dict() _lowerCamelCase = self.text_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
638
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } _lowerCamelCase = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub _lowerCamelCase = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase_ ( self , **A_ ) -> str: """simple docstring""" _lowerCamelCase = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> Optional[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> int: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = feature_extractor(A_ , return_tensors='''np''' ) _lowerCamelCase = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = '''This is a test string''' _lowerCamelCase = processor(text=A_ ) _lowerCamelCase = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase_ ( self , A_=(2, 10, 16) , A_=77 ) -> Optional[Any]: """simple docstring""" np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits(shape=(10, 16) , seed=13 ) _lowerCamelCase = processor.decode(A_ ) _lowerCamelCase = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: _lowerCamelCase = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: _lowerCamelCase = processor.batch_decode(A_ , A_ ) _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as p: _lowerCamelCase = decoder.decode_beams_batch(A_ , A_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 15 _lowerCamelCase = -20.0 _lowerCamelCase = -4.0 _lowerCamelCase = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] _lowerCamelCase = [d[0][2] for d in decoded_decoder_out] _lowerCamelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 2.0 _lowerCamelCase = 5.0 _lowerCamelCase = -20.0 _lowerCamelCase = True _lowerCamelCase = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = snapshot_download('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = processor_wavaveca(A_ , return_tensors='''np''' ) _lowerCamelCase = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor_wavaveca.batch_decode(A_ ) _lowerCamelCase = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = [d[key] for d in offsets] return retrieved_list def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits()[0] _lowerCamelCase = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" import torch _lowerCamelCase = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) _lowerCamelCase = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) _lowerCamelCase = iter(A_ ) _lowerCamelCase = next(A_ ) _lowerCamelCase = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) _lowerCamelCase = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train _lowerCamelCase = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): _lowerCamelCase = model(A_ ).logits.cpu().numpy() _lowerCamelCase = processor.decode(logits[0] , output_word_offsets=A_ ) _lowerCamelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate _lowerCamelCase = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] _lowerCamelCase = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off _lowerCamelCase = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) _lowerCamelCase = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
638
1
import math def __magic_name__( __UpperCAmelCase ) -> list: '''simple docstring''' _lowerCamelCase = [True] * n _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): _lowerCamelCase = i * 2 while index < n: _lowerCamelCase = False _lowerCamelCase = index + i _lowerCamelCase = [2] for i in range(3 , __UpperCAmelCase , 2 ): if is_prime[i]: primes.append(__UpperCAmelCase ) return primes def __magic_name__( __UpperCAmelCase = 9999_6666_3333 ) -> int: '''simple docstring''' _lowerCamelCase = math.floor(math.sqrt(__UpperCAmelCase ) ) + 100 _lowerCamelCase = prime_sieve(__UpperCAmelCase ) _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = primes[prime_index] while (last_prime**2) <= limit: _lowerCamelCase = primes[prime_index + 1] _lowerCamelCase = last_prime**2 _lowerCamelCase = next_prime**2 # Get numbers divisible by lps(current) _lowerCamelCase = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) _lowerCamelCase = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps _lowerCamelCase = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair _lowerCamelCase = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _lowerCamelCase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _lowerCamelCase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _lowerCamelCase = subset[i - 1][j] if arr[i - 1] <= j: _lowerCamelCase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
638
1
import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' random.seed(__UpperCAmelCase ) np.random.seed(__UpperCAmelCase ) torch.manual_seed(__UpperCAmelCase ) torch.cuda.manual_seed_all(__UpperCAmelCase ) # ^^ safe to call this function even if cuda is not available class UpperCamelCase : '''simple docstring''' def __init__( self , A_ , A_ = 0.9999 , A_ = 0.0 , A_ = 0 , A_ = False , A_ = 1.0 , A_ = 2 / 3 , A_ = None , A_ = None , **A_ , ) -> Optional[Any]: """simple docstring""" if isinstance(A_ , torch.nn.Module ): _lowerCamelCase = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage`''' , '''1.0.0''' , A_ , standard_warn=A_ , ) _lowerCamelCase = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _lowerCamelCase = True if kwargs.get('''max_value''' , A_ ) is not None: _lowerCamelCase = '''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('''max_value''' , '''1.0.0''' , A_ , standard_warn=A_ ) _lowerCamelCase = kwargs['''max_value'''] if kwargs.get('''min_value''' , A_ ) is not None: _lowerCamelCase = '''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('''min_value''' , '''1.0.0''' , A_ , standard_warn=A_ ) _lowerCamelCase = kwargs['''min_value'''] _lowerCamelCase = list(A_ ) _lowerCamelCase = [p.clone().detach() for p in parameters] if kwargs.get('''device''' , A_ ) is not None: _lowerCamelCase = '''The `device` argument is deprecated. Please use `to` instead.''' deprecate('''device''' , '''1.0.0''' , A_ , standard_warn=A_ ) self.to(device=kwargs['''device'''] ) _lowerCamelCase = None _lowerCamelCase = decay _lowerCamelCase = min_decay _lowerCamelCase = update_after_step _lowerCamelCase = use_ema_warmup _lowerCamelCase = inv_gamma _lowerCamelCase = power _lowerCamelCase = 0 _lowerCamelCase = None # set in `step()` _lowerCamelCase = model_cls _lowerCamelCase = model_config @classmethod def UpperCamelCase_ ( cls , A_ , A_ ) -> "EMAModel": """simple docstring""" _lowerCamelCase , _lowerCamelCase = model_cls.load_config(A_ , return_unused_kwargs=A_ ) _lowerCamelCase = model_cls.from_pretrained(A_ ) _lowerCamelCase = cls(model.parameters() , model_cls=A_ , model_config=model.config ) ema_model.load_state_dict(A_ ) return ema_model def UpperCamelCase_ ( self , A_ ) -> Optional[Any]: """simple docstring""" if self.model_cls is None: raise ValueError('''`save_pretrained` can only be used if `model_cls` was defined at __init__.''' ) if self.model_config is None: raise ValueError('''`save_pretrained` can only be used if `model_config` was defined at __init__.''' ) _lowerCamelCase = self.model_cls.from_config(self.model_config ) _lowerCamelCase = self.state_dict() state_dict.pop('''shadow_params''' , A_ ) model.register_to_config(**A_ ) self.copy_to(model.parameters() ) model.save_pretrained(A_ ) def UpperCamelCase_ ( self , A_ ) -> float: """simple docstring""" _lowerCamelCase = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _lowerCamelCase = 1 - (1 + step / self.inv_gamma) ** -self.power else: _lowerCamelCase = (1 + step) / (10 + step) _lowerCamelCase = min(A_ , self.decay ) # make sure decay is not smaller than min_decay _lowerCamelCase = max(A_ , self.min_decay ) return cur_decay_value @torch.no_grad() def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" if isinstance(A_ , torch.nn.Module ): _lowerCamelCase = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage.step`''' , '''1.0.0''' , A_ , standard_warn=A_ , ) _lowerCamelCase = parameters.parameters() _lowerCamelCase = list(A_ ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _lowerCamelCase = self.get_decay(self.optimization_step ) _lowerCamelCase = decay _lowerCamelCase = 1 - decay _lowerCamelCase = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , A_ ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _lowerCamelCase = deepspeed.zero.GatheredParameters(A_ , modifier_rank=A_ ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(A_ ) def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" _lowerCamelCase = list(A_ ) for s_param, param in zip(self.shadow_params , A_ ): param.data.copy_(s_param.to(param.device ).data ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> None: """simple docstring""" _lowerCamelCase = [ p.to(device=A_ , dtype=A_ ) if p.is_floating_point() else p.to(device=A_ ) for p in self.shadow_params ] def UpperCamelCase_ ( self ) -> dict: """simple docstring""" return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" _lowerCamelCase = [param.detach().cpu().clone() for param in parameters] def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" if self.temp_stored_params is None: raise RuntimeError('''This ExponentialMovingAverage has no `store()`ed weights ''' '''to `restore()`''' ) for c_param, param in zip(self.temp_stored_params , A_ ): param.data.copy_(c_param.data ) # Better memory-wise. _lowerCamelCase = None def UpperCamelCase_ ( self , A_ ) -> None: """simple docstring""" _lowerCamelCase = copy.deepcopy(A_ ) _lowerCamelCase = state_dict.get('''decay''' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('''Decay must be between 0 and 1''' ) _lowerCamelCase = state_dict.get('''min_decay''' , self.min_decay ) if not isinstance(self.min_decay , A_ ): raise ValueError('''Invalid min_decay''' ) _lowerCamelCase = state_dict.get('''optimization_step''' , self.optimization_step ) if not isinstance(self.optimization_step , A_ ): raise ValueError('''Invalid optimization_step''' ) _lowerCamelCase = state_dict.get('''update_after_step''' , self.update_after_step ) if not isinstance(self.update_after_step , A_ ): raise ValueError('''Invalid update_after_step''' ) _lowerCamelCase = state_dict.get('''use_ema_warmup''' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , A_ ): raise ValueError('''Invalid use_ema_warmup''' ) _lowerCamelCase = state_dict.get('''inv_gamma''' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('''Invalid inv_gamma''' ) _lowerCamelCase = state_dict.get('''power''' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('''Invalid power''' ) _lowerCamelCase = state_dict.get('''shadow_params''' , A_ ) if shadow_params is not None: _lowerCamelCase = shadow_params if not isinstance(self.shadow_params , A_ ): raise ValueError('''shadow_params must be a list''' ) if not all(isinstance(A_ , torch.Tensor ) for p in self.shadow_params ): raise ValueError('''shadow_params must all be Tensors''' )
638
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
1
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> None: """simple docstring""" warnings.warn( '''The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use GLPNImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
638
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 'roformer' def __init__( self , A_=5_00_00 , A_=None , A_=7_68 , A_=12 , A_=12 , A_=30_72 , A_="gelu" , A_=0.1 , A_=0.1 , A_=15_36 , A_=2 , A_=0.02 , A_=1E-1_2 , A_=0 , A_=False , A_=True , **A_ , ) -> Tuple: """simple docstring""" super().__init__(pad_token_id=A_ , **A_ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size if embedding_size is None else embedding_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = rotary_value _lowerCamelCase = use_cache class UpperCamelCase ( __lowercase ): '''simple docstring''' @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": _lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase = {0: '''batch''', 1: '''sequence'''} _lowerCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
638
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer snake_case__ = logging.get_logger(__name__) # pylint: disable=invalid-name snake_case__ = '\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu")\n\n >>> repo = "openai/shap-e-img2img"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png"\n >>> image = load_image(image_url).convert("RGB")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], "corgi_3d.gif")\n ```\n' @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , A_ , A_ , A_ , A_ , A_ , ) -> Dict: """simple docstring""" super().__init__() self.register_modules( prior=A_ , image_encoder=A_ , image_processor=A_ , scheduler=A_ , renderer=A_ , ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ , A_ , A_ ) -> List[Any]: """simple docstring""" if latents is None: _lowerCamelCase = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {shape}' ) _lowerCamelCase = latents.to(A_ ) _lowerCamelCase = latents * scheduler.init_noise_sigma return latents def UpperCamelCase_ ( self , A_=0 ) -> List[Any]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) _lowerCamelCase = torch.device(F'cuda:{gpu_id}' ) _lowerCamelCase = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder , '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(A_ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ , ) -> Optional[int]: """simple docstring""" if isinstance(A_ , A_ ) and isinstance(image[0] , torch.Tensor ): _lowerCamelCase = torch.cat(A_ , axis=0 ) if image[0].ndim == 4 else torch.stack(A_ , axis=0 ) if not isinstance(A_ , torch.Tensor ): _lowerCamelCase = self.image_processor(A_ , return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) _lowerCamelCase = image.to(dtype=self.image_encoder.dtype , device=A_ ) _lowerCamelCase = self.image_encoder(A_ )['''last_hidden_state'''] _lowerCamelCase = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 _lowerCamelCase = image_embeds.repeat_interleave(A_ , dim=0 ) if do_classifier_free_guidance: _lowerCamelCase = torch.zeros_like(A_ ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _lowerCamelCase = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self , A_ , A_ = 1 , A_ = 25 , A_ = None , A_ = None , A_ = 4.0 , A_ = 64 , A_ = "pil" , A_ = True , ) -> Any: """simple docstring""" if isinstance(A_ , PIL.Image.Image ): _lowerCamelCase = 1 elif isinstance(A_ , torch.Tensor ): _lowerCamelCase = image.shape[0] elif isinstance(A_ , A_ ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): _lowerCamelCase = len(A_ ) else: raise ValueError( F'`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(A_ )}' ) _lowerCamelCase = self._execution_device _lowerCamelCase = batch_size * num_images_per_prompt _lowerCamelCase = guidance_scale > 1.0 _lowerCamelCase = self._encode_image(A_ , A_ , A_ , A_ ) # prior self.scheduler.set_timesteps(A_ , device=A_ ) _lowerCamelCase = self.scheduler.timesteps _lowerCamelCase = self.prior.config.num_embeddings _lowerCamelCase = self.prior.config.embedding_dim _lowerCamelCase = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim _lowerCamelCase = latents.reshape(latents.shape[0] , A_ , A_ ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance _lowerCamelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents _lowerCamelCase = self.scheduler.scale_model_input(A_ , A_ ) _lowerCamelCase = self.prior( A_ , timestep=A_ , proj_embedding=A_ , ).predicted_image_embedding # remove the variance _lowerCamelCase , _lowerCamelCase = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: _lowerCamelCase , _lowerCamelCase = noise_pred.chunk(2 ) _lowerCamelCase = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) _lowerCamelCase = self.scheduler.step( A_ , timestep=A_ , sample=A_ , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=A_ ) _lowerCamelCase = [] for i, latent in enumerate(A_ ): print() _lowerCamelCase = self.renderer.decode( latent[None, :] , A_ , size=A_ , ray_batch_size=40_96 , n_coarse_samples=64 , n_fine_samples=1_28 , ) images.append(A_ ) _lowerCamelCase = torch.stack(A_ ) if output_type not in ["np", "pil"]: raise ValueError(F'Only the output types `pil` and `np` are supported not output_type={output_type}' ) _lowerCamelCase = images.cpu().numpy() if output_type == "pil": _lowerCamelCase = [self.numpy_to_pil(A_ ) for image in images] # Offload last model to CPU if hasattr(self , '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=A_ )
638
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ) -> None: """simple docstring""" _lowerCamelCase = device _lowerCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _lowerCamelCase = [0.48145466, 0.4578275, 0.40821073] _lowerCamelCase = [0.26862954, 0.26130258, 0.27577711] _lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCamelCase = torchvision.transforms.Resize(2_24 ) _lowerCamelCase = torchvision.transforms.CenterCrop(2_24 ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.resize(A_ ) _lowerCamelCase = self.center_crop(A_ ) _lowerCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer(text=A_ , **A_ ) _lowerCamelCase = self.preprocess_img(A_ ) _lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ) -> None: """simple docstring""" super().__init__() _lowerCamelCase = None _lowerCamelCase = device if device else get_device() if vqgan: _lowerCamelCase = vqgan else: _lowerCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _lowerCamelCase = clip else: _lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _lowerCamelCase = ProcessorGradientFlow(device=self.device ) _lowerCamelCase = iterations _lowerCamelCase = lr _lowerCamelCase = log _lowerCamelCase = make_grid _lowerCamelCase = return_val _lowerCamelCase = quantize _lowerCamelCase = self.vqgan.decoder.z_shape def UpperCamelCase_ ( self , A_=None , A_=None , A_=5 , A_=True ) -> Any: """simple docstring""" _lowerCamelCase = [] if output_path is None: _lowerCamelCase = '''./animation.gif''' if input_path is None: _lowerCamelCase = self.save_path _lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(A_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(A_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _lowerCamelCase = total_duration / len(A_ ) _lowerCamelCase = [frame_duration] * len(A_ ) if extend_frames: _lowerCamelCase = 1.5 _lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F'gif saved to {output_path}' ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> Union[str, Any]: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _lowerCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _lowerCamelCase = preprocess_vqgan(A_ ) _lowerCamelCase , *_lowerCamelCase = self.vqgan.encode(A_ ) return z def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.latent.detach().requires_grad_() _lowerCamelCase = base_latent + transform_vector if self.quantize: _lowerCamelCase , *_lowerCamelCase = self.vqgan.quantize(A_ ) else: _lowerCamelCase = trans_latent return self.vqgan.decode(A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_=None ) -> Any: """simple docstring""" _lowerCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors='''pt''' , padding=A_ ) _lowerCamelCase = self.clip(**A_ ) _lowerCamelCase = clip_outputs.logits_per_image if weights is not None: _lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , A_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , A_ , weights=neg_prompts['''weights'''] ) else: _lowerCamelCase = torch.tensor([1] , device=self.device ) _lowerCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCamelCase = self._add_vector(A_ ) _lowerCamelCase = loop_post_process(A_ ) _lowerCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print('''CLIP loss''' , A_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" wandb.init(reinit=A_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _lowerCamelCase = Image.open(A_ ) _lowerCamelCase = image.resize((2_56, 2_56) ) wandb.log('''Original Image''' , wandb.Image(A_ ) ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if not prompts: return [] _lowerCamelCase = [] _lowerCamelCase = [] if isinstance(A_ , A_ ): _lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _lowerCamelCase = prompt[0] _lowerCamelCase = float(prompt[1] ) elif ":" in prompt: _lowerCamelCase , _lowerCamelCase = prompt.split(''':''' ) _lowerCamelCase = float(A_ ) else: _lowerCamelCase = prompt _lowerCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ) -> str: """simple docstring""" if image_path: _lowerCamelCase = self._get_latent(A_ ) else: _lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCamelCase = self.process_prompts(A_ ) _lowerCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(A_ ) _lowerCamelCase = save_path _lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(A_ ) ) _lowerCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({'''Image''': wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
638
1
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): snake_case__ = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right snake_case__ = 12_8022 snake_case__ = 12_8028 @require_sentencepiece class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = MaMaaaTokenizer A_ = False A_ = False A_ = True def UpperCamelCase_ ( self ) -> str: """simple docstring""" super().setUp() _lowerCamelCase = ['''</s>''', '''<unk>''', '''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est''', '''\u0120''', '''<pad>'''] _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = Path(self.tmpdirname ) save_json(A_ , save_dir / VOCAB_FILES_NAMES['''vocab_file'''] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(A_ , save_dir / VOCAB_FILES_NAMES['''spm_file'''] ) _lowerCamelCase = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self , **A_ ) -> Union[str, Any]: """simple docstring""" return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , A_ ) -> List[str]: """simple docstring""" return ( "This is a test", "This is a test", ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = '''</s>''' _lowerCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ) , A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ) , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''</s>''' ) self.assertEqual(vocab_keys[1] , '''<unk>''' ) self.assertEqual(vocab_keys[-1] , '''<s>''' ) self.assertEqual(len(A_ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip('''Skip this test while all models are still to be uploaded.''' ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [2, 3, 4, 5, 6] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) _lowerCamelCase = tokenizer.convert_tokens_to_string(A_ ) self.assertEqual(A_ , '''This is a test''' ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # fmt: off _lowerCamelCase = {'''input_ids''': [[12_80_22, 11_01_08, 3_97, 11, 3_82_72, 22_47, 12_48_11, 2_85, 1_81_05, 15_86, 2_07, 7, 3_95_34, 44_28, 3_97, 10_19, 1_81_05, 15_86, 2_07, 7, 4_13_37, 1_67_86, 2_41, 7, 2_02_14, 17, 12_56_90, 1_03_98, 7, 4_43_78, 5_80_69, 6_83_42, 77_98, 73_43, 11, 2_99, 3_33_10, 4, 1_58, 3_73_50, 9_40_77, 45_69, 2_99, 3_33_10, 90, 4, 5_28_40, 2_90, 4, 3_12_70, 1_12, 2_99, 6_82, 4, 5_28_40, 3_99_53, 1_40_79, 1_93, 5_25_19, 9_08_94, 1_78_94, 12_06_97, 11, 4_04_45, 5_51, 17, 10_19, 5_25_19, 9_08_94, 1_77_56, 9_63, 11, 4_04_45, 4_80, 17, 97_92, 11_20, 51_73, 13_93, 62_40, 1_67_86, 2_41, 12_09_96, 28, 12_45, 13_93, 11_82_40, 1_11_23, 10_19, 9_36_12, 26_91, 1_06_18, 9_80_58, 12_04_09, 19_28, 2_79, 4, 4_06_83, 3_67, 1_78, 2_07, 10_19, 1_03, 10_31_21, 5_06, 6_52_96, 5, 2], [12_80_22, 2_12_17, 3_67, 1_17, 12_54_50, 1_28, 7_19, 7, 73_08, 40, 9_36_12, 1_26_69, 11_16, 1_67_04, 71, 1_77_85, 36_99, 1_55_92, 35, 1_44, 95_84, 2_41, 1_19_43, 7_13, 9_50, 7_99, 22_47, 8_84_27, 1_50, 1_49, 11_88_13, 12_07_06, 10_19, 10_69_06, 8_15_18, 28, 12_24, 2_27_99, 3_97, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [12_80_22, 16_58, 12_33_11, 51_55, 55_78, 47_22, 2_79, 1_49_47, 23_66, 11_20, 11_97, 14, 13_48, 92_32, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A_ , model_name='''facebook/m2m100_418M''' , revision='''c168bae485c864188cf9aa0e4108b0b6934dc91e''' , ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' A_ = 'facebook/m2m100_418M' A_ = [ 'In my opinion, there are two levels of response from the French government.', 'NSA Affair Emphasizes Complete Lack of Debate on Intelligence', ] A_ = [ 'Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.', 'L\'affaire NSA souligne l\'absence totale de débat sur le renseignement', ] # fmt: off A_ = [EN_CODE, 593, 1_949, 115_781, 4, 71_586, 4_234, 60_633, 126_233, 432, 123_808, 15_592, 1_197, 117_132, 120_618, 5, 2] @classmethod def UpperCamelCase_ ( cls ) -> str: """simple docstring""" _lowerCamelCase = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en''' , tgt_lang='''fr''' ) _lowerCamelCase = 1 return cls def UpperCamelCase_ ( self ) -> int: """simple docstring""" self.assertEqual(self.tokenizer.get_lang_id('''ar''' ) , 12_80_06 ) self.assertEqual(self.tokenizer.get_lang_id('''en''' ) , 12_80_22 ) self.assertEqual(self.tokenizer.get_lang_id('''ro''' ) , 12_80_76 ) self.assertEqual(self.tokenizer.get_lang_id('''mr''' ) , 12_80_63 ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.tokenizer.get_vocab() self.assertEqual(len(A_ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab['''<unk>'''] , 3 ) self.assertIn(self.tokenizer.get_lang_token('''en''' ) , A_ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = '''en''' _lowerCamelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" self.assertIn(A_ , self.tokenizer.all_special_ids ) # fmt: off _lowerCamelCase = [FR_CODE, 53_64, 82, 86_42, 4, 2_94, 47, 8, 1_40_28, 1_36, 32_86, 97_06, 6, 9_07_97, 6, 14_40_12, 1_62, 8_81_28, 3_00_61, 5, 2] # fmt: on _lowerCamelCase = self.tokenizer.decode(A_ , skip_special_tokens=A_ ) _lowerCamelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ ) self.assertNotIn(self.tokenizer.eos_token , A_ ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(A_ ) _lowerCamelCase = MaMaaaTokenizer.from_pretrained(A_ ) self.assertDictEqual(new_tok.lang_token_to_id , A_ ) @require_torch def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = '''en''' _lowerCamelCase = '''fr''' _lowerCamelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=A_ , return_tensors='''pt''' ) _lowerCamelCase = shift_tokens_right( batch['''labels'''] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _lowerCamelCase = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = '''mr''' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('''mr''' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _lowerCamelCase = '''zh''' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('''zh''' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = '''mr''' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('''mr''' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _lowerCamelCase = '''zh''' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('''zh''' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer._build_translation_inputs('''A test''' , return_tensors='''pt''' , src_lang='''en''' , tgt_lang='''ar''' ) self.assertEqual( nested_simplify(A_ ) , { # en_XX, A, test, EOS '''input_ids''': [[12_80_22, 58, 41_83, 2]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 12_80_06, } , )
638
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case__ = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class UpperCamelCase ( __lowercase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A_ , '''tf_padding''' ) ) self.parent.assertTrue(hasattr(A_ , '''depth_multiplier''' ) ) class UpperCamelCase : '''simple docstring''' def __init__( self , A_ , A_=13 , A_=3 , A_=32 , A_=0.25 , A_=8 , A_=8 , A_=6 , A_=32 , A_=True , A_=True , A_=True , A_="relu6" , A_=12_80 , A_=0.1 , A_=0.02 , A_=True , A_=True , A_=10 , A_=None , ) -> Any: """simple docstring""" _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = image_size _lowerCamelCase = depth_multiplier _lowerCamelCase = depth_divisible_by _lowerCamelCase = min_depth _lowerCamelCase = expand_ratio _lowerCamelCase = tf_padding _lowerCamelCase = output_stride _lowerCamelCase = first_layer_is_expansion _lowerCamelCase = finegrained_output _lowerCamelCase = hidden_act _lowerCamelCase = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) _lowerCamelCase = classifier_dropout_prob _lowerCamelCase = use_labels _lowerCamelCase = is_training _lowerCamelCase = num_labels _lowerCamelCase = initializer_range _lowerCamelCase = scope def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels, pixel_labels def UpperCamelCase_ ( self ) -> int: """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = MobileNetVaModel(config=A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.num_labels _lowerCamelCase = MobileNetVaForImageClassification(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , A_ , A_ , A_ , A_ ) -> List[str]: """simple docstring""" _lowerCamelCase = self.num_labels _lowerCamelCase = MobileNetVaForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() _lowerCamelCase = model(A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCamelCase = model(A_ , labels=A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) A_ = ( { 'feature-extraction': MobileNetVaModel, 'image-classification': MobileNetVaForImageClassification, 'image-segmentation': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) A_ = False A_ = False A_ = False A_ = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = MobileNetVaModelTester(self ) _lowerCamelCase = MobileNetVaConfigTester(self , config_class=A_ , has_text_modality=A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''MobileNetV2 does not use inputs_embeds''' ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass @unittest.skip(reason='''MobileNetV2 does not support input and output embeddings''' ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" pass @unittest.skip(reason='''MobileNetV2 does not output attentions''' ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(A_ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" def check_hidden_states_output(A_ , A_ , A_ ): _lowerCamelCase = model_class(A_ ) model.to(A_ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(A_ , A_ ) ) _lowerCamelCase = outputs.hidden_states _lowerCamelCase = 16 self.assertEqual(len(A_ ) , A_ ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = True check_hidden_states_output(A_ , A_ , A_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(A_ , A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = MobileNetVaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def __magic_name__( ) -> str: '''simple docstring''' _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained('''google/mobilenet_v2_1.0_224''' ) if is_vision_available() else None ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = MobileNetVaForImageClassification.from_pretrained('''google/mobilenet_v2_1.0_224''' ).to(A_ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=A_ , return_tensors='''pt''' ).to(A_ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**A_ ) # verify the logits _lowerCamelCase = torch.Size((1, 10_01) ) self.assertEqual(outputs.logits.shape , A_ ) _lowerCamelCase = torch.tensor([0.2445, -1.1993, 0.1905] ).to(A_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A_ , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = MobileNetVaForSemanticSegmentation.from_pretrained('''google/deeplabv3_mobilenet_v2_1.0_513''' ) _lowerCamelCase = model.to(A_ ) _lowerCamelCase = MobileNetVaImageProcessor.from_pretrained('''google/deeplabv3_mobilenet_v2_1.0_513''' ) _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=A_ , return_tensors='''pt''' ).to(A_ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**A_ ) _lowerCamelCase = outputs.logits # verify the logits _lowerCamelCase = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , A_ ) _lowerCamelCase = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=A_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A_ , atol=1E-4 ) )
638
import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='%(message)s') def __magic_name__( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size, 1) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.mean(1 ) # Centralize the data of class i _lowerCamelCase = data - column_reshape(__UpperCAmelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(__UpperCAmelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = features.mean(1 ) _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.shape[1] _lowerCamelCase = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' if features.any(): _lowerCamelCase = features.mean(1 ) # Center the dataset _lowerCamelCase = features - np.reshape(__UpperCAmelCase , (data_mean.size, 1) ) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) / features.shape[1] _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(__UpperCAmelCase ) # Take all the columns in the reverse order (-1), and then takes only the first _lowerCamelCase = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space _lowerCamelCase = np.dot(filtered_eigenvectors.T , __UpperCAmelCase ) logging.info('''Principal Component Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' assert classes > dimensions # Check if features have been already loaded if features.any: _lowerCamelCase , _lowerCamelCase = eigh( covariance_between_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , covariance_within_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , ) _lowerCamelCase = eigenvectors[:, ::-1][:, :dimensions] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = np.linalg.svd(__UpperCAmelCase ) _lowerCamelCase = svd_matrix[:, 0:dimensions] _lowerCamelCase = np.dot(filtered_svd_matrix.T , __UpperCAmelCase ) logging.info('''Linear Discriminant Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) _lowerCamelCase = np.array([0, 0, 0, 1, 1] ) _lowerCamelCase = 2 _lowerCamelCase = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = linear_discriminant_analysis( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if isinstance(__UpperCAmelCase , np.ndarray ): raise AssertionError( '''Did not raise AssertionError for dimensions > classes''' ) assert error_info.type is AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) _lowerCamelCase = 2 _lowerCamelCase = np.array([[6.9_2_8_2_0_3_2_3, 8.6_6_0_2_5_4_0_4, 1_0.3_9_2_3_0_4_8_5], [3.0, 3.0, 3.0]] ) with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = principal_component_analysis(__UpperCAmelCase , __UpperCAmelCase ) if not np.allclose(__UpperCAmelCase , __UpperCAmelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
638
1
import socket def __magic_name__( ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase = socket.gethostname() _lowerCamelCase = 1_2312 sock.connect((host, port) ) sock.send(B'''Hello server!''' ) with open('''Received_file''' , '''wb''' ) as out_file: print('''File opened''' ) print('''Receiving data...''' ) while True: _lowerCamelCase = sock.recv(1024 ) if not data: break out_file.write(__UpperCAmelCase ) print('''Successfully received the file''' ) sock.close() print('''Connection closed''' ) if __name__ == "__main__": main()
638
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['vqvae'] def __init__( self , A_ , A_ , A_ , A_ , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules(unet=A_ , scheduler=A_ , mel=A_ , vqvae=A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 50 if isinstance(self.scheduler , A_ ) else 10_00 @torch.no_grad() def __call__( self , A_ = 1 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = None , A_ = 0 , A_ = 0 , A_ = None , A_ = 0 , A_ = None , A_ = None , A_=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: """simple docstring""" _lowerCamelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(A_ ) _lowerCamelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: _lowerCamelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: _lowerCamelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=A_ , device=self.device , ) _lowerCamelCase = noise _lowerCamelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A_ , A_ ) _lowerCamelCase = self.mel.audio_slice_to_image(A_ ) _lowerCamelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) _lowerCamelCase = (input_image / 2_55) * 2 - 1 _lowerCamelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: _lowerCamelCase = self.vqvae.encode(torch.unsqueeze(A_ , 0 ) ).latent_dist.sample( generator=A_ )[0] _lowerCamelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , self.scheduler.timesteps[start_step - 1] ) _lowerCamelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) _lowerCamelCase = int(mask_start_secs * pixels_per_second ) _lowerCamelCase = int(mask_end_secs * pixels_per_second ) _lowerCamelCase = self.scheduler.add_noise(A_ , A_ , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , A_ ): _lowerCamelCase = self.unet(A_ , A_ , A_ )['''sample'''] else: _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] if isinstance(self.scheduler , A_ ): _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , eta=A_ , generator=A_ , )['''prev_sample'''] else: _lowerCamelCase = self.scheduler.step( model_output=A_ , timestep=A_ , sample=A_ , generator=A_ , )['''prev_sample'''] if mask is not None: if mask_start > 0: _lowerCamelCase = mask[:, step, :, :mask_start] if mask_end > 0: _lowerCamelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance _lowerCamelCase = 1 / self.vqvae.config.scaling_factor * images _lowerCamelCase = self.vqvae.decode(A_ )['''sample'''] _lowerCamelCase = (images / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() _lowerCamelCase = (images * 2_55).round().astype('''uint8''' ) _lowerCamelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A_ , mode='''RGB''' ).convert('''L''' ) for _ in images) ) _lowerCamelCase = [self.mel.image_to_audio(A_ ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A_ )[:, np.newaxis, :] ) , **ImagePipelineOutput(A_ ) ) @torch.no_grad() def UpperCamelCase_ ( self , A_ , A_ = 50 ) -> np.ndarray: """simple docstring""" assert isinstance(self.scheduler , A_ ) self.scheduler.set_timesteps(A_ ) _lowerCamelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) _lowerCamelCase = (sample / 2_55) * 2 - 1 _lowerCamelCase = torch.Tensor(A_ ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): _lowerCamelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps _lowerCamelCase = self.scheduler.alphas_cumprod[t] _lowerCamelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) _lowerCamelCase = 1 - alpha_prod_t _lowerCamelCase = self.unet(A_ , A_ )['''sample'''] _lowerCamelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output _lowerCamelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) _lowerCamelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def UpperCamelCase_ ( A_ , A_ , A_ ) -> torch.Tensor: """simple docstring""" _lowerCamelCase = acos(torch.dot(torch.flatten(A_ ) , torch.flatten(A_ ) ) / torch.norm(A_ ) / torch.norm(A_ ) ) return sin((1 - alpha) * theta ) * xa / sin(A_ ) + sin(alpha * theta ) * xa / sin(A_ )
638
1
import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate snake_case__ = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow('', '|', '|'), datarow=DataRow('', '|', '|'), padding=1, with_header_hide=None, ) snake_case__ = [] snake_case__ = [] snake_case__ = {'type': 'section', 'text': {'type': 'plain_text', 'text': 'No failed tests! 🤗', 'emoji': True}} snake_case__ = [ { 'type': 'header', 'text': { 'type': 'plain_text', 'text': f'''🤗 Accelerate nightly {os.environ.get("TEST_TYPE", "")} test results''', 'emoji': True, }, } ] snake_case__ = 0 for log in Path().glob('*.log'): snake_case__ = 0 with open(log, 'r') as f: for line in f: snake_case__ = json.loads(line) if line.get('nodeid', '') != "": snake_case__ = line['nodeid'] if line.get('duration', None) is not None: snake_case__ = f'''{line["duration"]:.4f}''' if line.get('outcome', '') == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split('_')[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) snake_case__ = [] log.unlink() snake_case__ = '' snake_case__ = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += f"*{name[1:]}: {num_failed} failed test*\n" else: message += f"*{name[1:]}: {num_failed} failed tests*\n" snake_case__ = [] snake_case__ = {} for test in failed_tests: snake_case__ = test[0].split('::') snake_case__ = data[0].split('/')[-1] if data[0] not in filesafailed: snake_case__ = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) snake_case__ = [test[0] for test in failed_table] snake_case__ = list(set(files)) # Count number of instances in failed_tests snake_case__ = [] for file in individual_files: table.append([file, len(filesafailed[file])]) snake_case__ = tabulate( table, headers=['Test Location', 'Num Failed'], tablefmt=hf_table_format, stralign='right', ) message += f"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3000: snake_case__ = 'Too many failed tests, please see the full report in the Action results.' snake_case__ = len(err) + 10 snake_case__ = message[: 3000 - offset] + f'''\n...\n```\n{err}''' print(f'''### {message}''') else: snake_case__ = 'No failed tests! 🤗' print(f'''## {message}''') payload.append(no_error_payload) if os.environ.get('TEST_TYPE', '') != "": from slack_sdk import WebClient snake_case__ = WebClient(token=os.environ['SLACK_API_TOKEN']) if message != "No failed tests! 🤗": snake_case__ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': message, }, } payload.append(md_report) snake_case__ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': '*For more details:*', }, 'accessory': { 'type': 'button', 'text': { 'type': 'plain_text', 'text': 'Check Action results', 'emoji': True, }, 'url': f'''https://github.com/{os.environ["GITHUB_REPOSITORY"]}/actions/runs/{os.environ["GITHUB_RUN_ID"]}''', }, } payload.append(action_button) snake_case__ = { 'type': 'context', 'elements': [ { 'type': 'plain_text', 'text': f'''Nightly {os.environ.get("TEST_TYPE")} test results for {date.today()}''', } ], } payload.append(date_report) snake_case__ = client.chat_postMessage(channel='#accelerate-ci-daily', text=message, blocks=payload) snake_case__ = response.data['ts'] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name snake_case__ = '' for i, row in enumerate(test_failures): if row[0] != test_class: snake_case__ = row[0] else: snake_case__ = '' snake_case__ = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': f'''Test location: {test_location}\n```\n{tabulate(test_failures, headers=["Class", "Test"], tablefmt=hf_table_format, stralign="right")}\n```''', }, } client.chat_postMessage( channel='#accelerate-ci-daily', thread_ts=ts, blocks=[payload], )
638
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowercase ) , 'Tatoeba directory does not exist.' ) class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=A_ ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase , _lowerCamelCase = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=A_ ) assert mmeta["long_pair"] == "heb-eng"
638
1
def __magic_name__( __UpperCAmelCase ) -> int: # noqa: E741 '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = 0 _lowerCamelCase = [0] * n _lowerCamelCase = [False] * n _lowerCamelCase = [False] * n def dfs(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): if parent == root: out_edge_count += 1 _lowerCamelCase = True _lowerCamelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: _lowerCamelCase = dfs(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: _lowerCamelCase = True # AP found via cycle if at == low[to]: _lowerCamelCase = True else: _lowerCamelCase = min(low[at] , __UpperCAmelCase ) return out_edge_count for i in range(__UpperCAmelCase ): if not visited[i]: _lowerCamelCase = 0 _lowerCamelCase = dfs(__UpperCAmelCase , __UpperCAmelCase , -1 , __UpperCAmelCase ) _lowerCamelCase = out_edge_count > 1 for x in range(len(__UpperCAmelCase ) ): if is_art[x] is True: print(__UpperCAmelCase ) # Adjacency list of graph snake_case__ = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
638
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' _lowerCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _lowerCamelCase = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase = in_proj_bias[: config.hidden_size] _lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase = in_proj_bias[-config.hidden_size :] def __magic_name__( __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True ) -> str: '''simple docstring''' _lowerCamelCase = ViTConfig() # patch_size if model_name[-1] == "8": _lowerCamelCase = 8 # set labels if required if not base_model: _lowerCamelCase = 1000 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _lowerCamelCase = 384 _lowerCamelCase = 1536 _lowerCamelCase = 12 _lowerCamelCase = 6 # load original model from torch hub _lowerCamelCase = torch.hub.load('''facebookresearch/dino:main''' , __UpperCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase = original_model.state_dict() if base_model: remove_classification_head_(__UpperCAmelCase ) _lowerCamelCase = create_rename_keys(__UpperCAmelCase , base_model=__UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # load HuggingFace model if base_model: _lowerCamelCase = ViTModel(__UpperCAmelCase , add_pooling_layer=__UpperCAmelCase ).eval() else: _lowerCamelCase = ViTForImageClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _lowerCamelCase = ViTImageProcessor() _lowerCamelCase = image_processor(images=prepare_img() , return_tensors='''pt''' ) _lowerCamelCase = encoding['''pixel_values'''] _lowerCamelCase = model(__UpperCAmelCase ) if base_model: _lowerCamelCase = original_model(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _lowerCamelCase = original_model(__UpperCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) snake_case__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
638
1
from __future__ import annotations from dataclasses import dataclass @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = None A_ = None def __magic_name__( __UpperCAmelCase ) -> bool: '''simple docstring''' def is_valid_tree(__UpperCAmelCase ) -> bool: if node is None: return True if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return False try: float(node.data ) except (TypeError, ValueError): return False return is_valid_tree(node.left ) and is_valid_tree(node.right ) if not is_valid_tree(__UpperCAmelCase ): raise ValueError( '''Each node should be type of TreeNode and data should be float.''' ) def is_binary_search_tree_recursive_check( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: if node is None: return True return ( left_bound < node.data < right_bound and is_binary_search_tree_recursive_check(node.left , __UpperCAmelCase , node.data ) and is_binary_search_tree_recursive_check( node.right , node.data , __UpperCAmelCase ) ) return is_binary_search_tree_recursive_check(__UpperCAmelCase , -float('''inf''' ) , float('''inf''' ) ) if __name__ == "__main__": import doctest doctest.testmod()
638
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } snake_case__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _lowerCamelCase = '''lm_head''' _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ) if weight_type is not None: _lowerCamelCase = getattr(__UpperCAmelCase , __UpperCAmelCase ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): _lowerCamelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(__UpperCAmelCase )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , __UpperCAmelCase ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) continue if not is_used: unused_weights.append(__UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__UpperCAmelCase ) @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True ) -> Union[str, Any]: '''simple docstring''' if config_path is not None: _lowerCamelCase = UniSpeechConfig.from_pretrained(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _lowerCamelCase = Dictionary.load_from_json(__UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _lowerCamelCase = target_dict.pad_index _lowerCamelCase = target_dict.bos_index _lowerCamelCase = target_dict.eos_index _lowerCamelCase = len(target_dict.symbols ) _lowerCamelCase = os.path.join(__UpperCAmelCase , '''vocab.json''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__UpperCAmelCase ) ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) _lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _lowerCamelCase = 42 _lowerCamelCase = 43 with open(__UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = WavaVecaPhonemeCTCTokenizer( __UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__UpperCAmelCase , ) _lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False _lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) _lowerCamelCase = WavaVecaProcessor(feature_extractor=__UpperCAmelCase , tokenizer=__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) _lowerCamelCase = UniSpeechForCTC(__UpperCAmelCase ) else: _lowerCamelCase = UniSpeechForPreTraining(__UpperCAmelCase ) if is_finetuned: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _lowerCamelCase = model[0].eval() recursively_load_weights(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) hf_unispeech.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
638
1
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class UpperCamelCase ( __lowercase ): '''simple docstring''' @require_torch def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _lowerCamelCase = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _lowerCamelCase = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _lowerCamelCase = ''' import socket def offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _lowerCamelCase = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task='''fill-mask''' , model=A_ ) # baseline - just load from_pretrained with normal network _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _lowerCamelCase = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _lowerCamelCase = '''1''' _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # python one-liner segments # this must be loaded before socket.socket is monkey-patched _lowerCamelCase = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _lowerCamelCase = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _lowerCamelCase = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _lowerCamelCase = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task='''fill-mask''' , model=A_ ) # baseline - just load from_pretrained with normal network _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _lowerCamelCase = self.get_env() _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _lowerCamelCase = ''' from transformers import BertConfig, BertModel, BertTokenizer ''' _lowerCamelCase = ''' mname = "hf-internal-testing/tiny-random-bert-sharded" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print("success") ''' _lowerCamelCase = ''' import socket def offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled") socket.socket = offline_socket ''' # baseline - just load from_pretrained with normal network _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _lowerCamelCase = self.get_env() _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # next emulate no network _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _lowerCamelCase = '''1''' _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = ''' from transformers import pipeline ''' _lowerCamelCase = ''' mname = "hf-internal-testing/tiny-random-bert" pipe = pipeline(model=mname) ''' _lowerCamelCase = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled") socket.socket = offline_socket ''' _lowerCamelCase = self.get_env() _lowerCamelCase = '''1''' _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( '''You cannot infer task automatically within `pipeline` when using offline mode''' , result.stderr.decode().replace('''\n''' , '''''' ) , ) @require_torch def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = ''' from transformers import AutoModel ''' _lowerCamelCase = ''' mname = "hf-internal-testing/test_dynamic_model" AutoModel.from_pretrained(mname, trust_remote_code=True) print("success") ''' # baseline - just load from_pretrained with normal network _lowerCamelCase = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _lowerCamelCase = self.get_env() _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _lowerCamelCase = '''1''' _lowerCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() )
638
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' def __init__( self , *A_ , **A_ ) -> None: """simple docstring""" warnings.warn( '''The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DPTImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
638
1
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case__ = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = ReformerTokenizer A_ = ReformerTokenizerFast A_ = True A_ = False A_ = True def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setUp() _lowerCamelCase = ReformerTokenizer(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = '''<s>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ) , A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ) , A_ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(A_ ) , 10_00 ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(A_ ) _lowerCamelCase = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) _lowerCamelCase = tokenizer.encode(A_ , add_special_tokens=A_ ) _lowerCamelCase = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(A_ ) _lowerCamelCase = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self , A_=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) # Simple input _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _lowerCamelCase = ('''This is a simple input''', '''This is a pair''') _lowerCamelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='''max_length''' ) # Simple input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='''max_length''' ) # Simple input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='''max_length''' , ) # Pair input self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='''max_length''' ) # Pair input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='''max_length''' ) # Pair input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='''max_length''' , ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = ReformerTokenizer(A_ , keep_accents=A_ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [2_85, 46, 10, 1_70, 3_82] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [1_26, 32, 2_62, 1_52, 38, 72, 2_87] self.assertListEqual(A_ , self.big_tokenizer.encode(A_ ) ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 1_08, 2_65, 24, 1_11, 4, 2_58, 1_56, 35, 28, 2_75, 3, 2_59, 2_97, 2_60, 84, 4, 35, 1_10, 44, 8, 2_59, 91, 2_68, 21, 11, 2_09, 2_74, 1_09, 2_66, 2_77, 1_17, 86, 93, 3_15, 2_58, 2_78, 2_58, 2_77, 2_58, 0, 2_58, 2_88, 2_58, 3_19, 2_58, 0, 2_58, 0, 2_58, 0, 2_58, 0, 2_58, 2_87, 2_58, 3_15, 2_58, 2_89, 2_58, 2_78, 99, 2_69, 2_66, 2_62, 8, 2_59, 2_41, 4, 2_17, 2_30, 2_68, 2_66, 55, 1_68, 1_06, 75, 1_93, 2_66, 2_23, 27, 49, 26, 2_82, 25, 2_64, 2_99, 19, 26, 0, 2_58, 2_77, 1_17, 86, 93, 1_76, 1_83, 2_70, 11, 2_62, 42, 61, 2_65, ] self.assertListEqual(A_ , self.big_tokenizer.encode(A_ ) ) @require_torch @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence _lowerCamelCase = list(self.big_tokenizer.get_vocab().keys() )[:10] _lowerCamelCase = ''' '''.join(A_ ) _lowerCamelCase = self.big_tokenizer.encode_plus(A_ , return_tensors='''pt''' ) _lowerCamelCase = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) _lowerCamelCase = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _lowerCamelCase = encoded_sequence['''input_ids'''].shape _lowerCamelCase = ReformerModel(A_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**A_ ) model(**A_ ) @slow def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # fmt: off _lowerCamelCase = {'''input_ids''': [[1_08, 2_65, 24, 1_11, 4, 2_58, 1_56, 7, 51, 2_79, 58, 7, 76, 25, 69, 2_78], [1_40, 2_43, 2_64, 1_34, 17, 2_67, 77, 2_63, 22, 2_62, 2_97, 2_58, 3_04, 1_77, 2_79, 2_66, 14, 89, 13, 35, 2_61, 2_99, 2_72, 1_37, 2_75, 2_78]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _lowerCamelCase = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=A_ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=A_ , sequences=A_ , )
638
import argparse import json import subprocess def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = ( F'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) _lowerCamelCase = subprocess.run(__UpperCAmelCase , shell=__UpperCAmelCase , stdout=subprocess.PIPE ) _lowerCamelCase = output.stdout.decode('''utf-8''' ) _lowerCamelCase = json.loads(__UpperCAmelCase ) _lowerCamelCase = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__UpperCAmelCase ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) ) if len(__UpperCAmelCase ) > 0: _lowerCamelCase = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' return values.split(''',''' ) snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) snake_case__ = parser.parse_args() get_runner_status(args.target_runners, args.token)
638
1
def __magic_name__( __UpperCAmelCase ) -> list[list]: '''simple docstring''' _lowerCamelCase = current_set.copy() for row_index, row in enumerate(__UpperCAmelCase ): _lowerCamelCase = row[0] for column_index, column in enumerate(__UpperCAmelCase ): if magnitude == 0: _lowerCamelCase = column continue _lowerCamelCase = column / magnitude # Subtract to cancel term _lowerCamelCase = current_set[0] _lowerCamelCase = [first_row] _lowerCamelCase = current_set[1::] for row in current_set: _lowerCamelCase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(__UpperCAmelCase ) continue for column_index in range(len(__UpperCAmelCase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(__UpperCAmelCase ) # Create next recursion iteration set if len(final_set[0] ) != 3: _lowerCamelCase = final_set[0] _lowerCamelCase = [] _lowerCamelCase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) _lowerCamelCase = simplify(__UpperCAmelCase ) for i in range(len(__UpperCAmelCase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , __UpperCAmelCase ) _lowerCamelCase = resultant return final_set def __magic_name__( __UpperCAmelCase ) -> list: '''simple docstring''' if len(__UpperCAmelCase ) == 0: raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) _lowerCamelCase = len(__UpperCAmelCase ) + 1 if any(len(__UpperCAmelCase ) != _length for item in equations ): raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) for row in equations: if any(not isinstance(__UpperCAmelCase , (int, float) ) for column in row ): raise ValueError('''solve_simultaneous() requires lists of integers''' ) if len(__UpperCAmelCase ) == 1: return [equations[0][-1] / equations[0][0]] _lowerCamelCase = equations.copy() if any(0 in row for row in data_set ): _lowerCamelCase = data_set.copy() _lowerCamelCase = [] for row_index, row in enumerate(__UpperCAmelCase ): if 0 not in row: _lowerCamelCase = data_set.pop(__UpperCAmelCase ) break if not full_row: raise ValueError('''solve_simultaneous() requires at least 1 full equation''' ) data_set.insert(0 , __UpperCAmelCase ) _lowerCamelCase = data_set.copy() _lowerCamelCase = simplify(__UpperCAmelCase ) _lowerCamelCase = simplified[::-1] _lowerCamelCase = [] for row in simplified: _lowerCamelCase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue _lowerCamelCase = row.copy()[: len(__UpperCAmelCase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(__UpperCAmelCase ) == 0: solutions.append(0 ) continue _lowerCamelCase = temp_row[1::] _lowerCamelCase = temp_row[::-1] for column_index, column in enumerate(__UpperCAmelCase ): current_solution -= column * solutions[column_index] solutions.append(__UpperCAmelCase ) _lowerCamelCase = [] for item in solutions: final.append(float(round(__UpperCAmelCase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() snake_case__ = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
638
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
638
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case__ = logging.get_logger(__name__) snake_case__ = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' A_ = 'bit' A_ = ['preactivation', 'bottleneck'] A_ = ['SAME', 'VALID'] def __init__( self , A_=3 , A_=64 , A_=[2_56, 5_12, 10_24, 20_48] , A_=[3, 4, 6, 3] , A_="preactivation" , A_="relu" , A_=None , A_=32 , A_=0.0 , A_=False , A_=32 , A_=1 , A_=None , A_=None , **A_ , ) -> List[Any]: """simple docstring""" super().__init__(**A_ ) if layer_type not in self.layer_types: raise ValueError(F'layer_type={layer_type} is not one of {",".join(self.layer_types )}' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: _lowerCamelCase = global_padding.upper() else: raise ValueError(F'Padding strategy {global_padding} not supported' ) _lowerCamelCase = num_channels _lowerCamelCase = embedding_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = layer_type _lowerCamelCase = hidden_act _lowerCamelCase = global_padding _lowerCamelCase = num_groups _lowerCamelCase = drop_path_rate _lowerCamelCase = embedding_dynamic_padding _lowerCamelCase = output_stride _lowerCamelCase = width_factor _lowerCamelCase = ['''stem'''] + [F'stage{idx}' for idx in range(1 , len(A_ ) + 1 )] _lowerCamelCase , _lowerCamelCase = get_aligned_output_features_output_indices( out_features=A_ , out_indices=A_ , stage_names=self.stage_names )
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case__ = { 'configuration_bloom': ['BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BloomConfig', 'BloomOnnxConfig'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['BloomTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST', 'BloomForCausalLM', 'BloomModel', 'BloomPreTrainedModel', 'BloomForSequenceClassification', 'BloomForTokenClassification', 'BloomForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaTokenizerFast, ) def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = model.config _lowerCamelCase = DonutSwinConfig( image_size=original_config.input_size , patch_size=4 , depths=original_config.encoder_layer , num_heads=[4, 8, 16, 32] , window_size=original_config.window_size , embed_dim=128 , ) _lowerCamelCase = MBartConfig( is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , add_cross_attention=__UpperCAmelCase , decoder_layers=original_config.decoder_layer , max_position_embeddings=original_config.max_position_embeddings , vocab_size=len( model.decoder.tokenizer ) , scale_embedding=__UpperCAmelCase , add_final_layer_norm=__UpperCAmelCase , ) return encoder_config, decoder_config def __magic_name__( __UpperCAmelCase ) -> Tuple: '''simple docstring''' if "encoder.model" in name: _lowerCamelCase = name.replace('''encoder.model''' , '''encoder''' ) if "decoder.model" in name: _lowerCamelCase = name.replace('''decoder.model''' , '''decoder''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if name.startswith('''encoder''' ): if "layers" in name: _lowerCamelCase = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "mask" not in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": _lowerCamelCase = '''encoder.layernorm.weight''' if name == "encoder.norm.bias": _lowerCamelCase = '''encoder.layernorm.bias''' return name def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(__UpperCAmelCase ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = int(key_split[5] ) _lowerCamelCase = model.encoder.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] elif "attn_mask" in key or key in ["encoder.model.norm.weight", "encoder.model.norm.bias"]: # HuggingFace implementation doesn't use attn_mask buffer # and model doesn't use final LayerNorms for the encoder pass else: _lowerCamelCase = val return orig_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ) -> int: '''simple docstring''' _lowerCamelCase = DonutModel.from_pretrained(__UpperCAmelCase ).eval() # load HuggingFace model _lowerCamelCase , _lowerCamelCase = get_configs(__UpperCAmelCase ) _lowerCamelCase = DonutSwinModel(__UpperCAmelCase ) _lowerCamelCase = MBartForCausalLM(__UpperCAmelCase ) _lowerCamelCase = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() _lowerCamelCase = original_model.state_dict() _lowerCamelCase = convert_state_dict(__UpperCAmelCase , __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) # verify results on scanned document _lowerCamelCase = load_dataset('''hf-internal-testing/example-documents''' ) _lowerCamelCase = dataset['''test'''][0]['''image'''].convert('''RGB''' ) _lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained(__UpperCAmelCase , from_slow=__UpperCAmelCase ) _lowerCamelCase = DonutImageProcessor( do_align_long_axis=original_model.config.align_long_axis , size=original_model.config.input_size[::-1] ) _lowerCamelCase = DonutProcessor(__UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values if model_name == "naver-clova-ix/donut-base-finetuned-docvqa": _lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' _lowerCamelCase = '''When is the coffee break?''' _lowerCamelCase = task_prompt.replace('''{user_input}''' , __UpperCAmelCase ) elif model_name == "naver-clova-ix/donut-base-finetuned-rvlcdip": _lowerCamelCase = '''<s_rvlcdip>''' elif model_name in [ "naver-clova-ix/donut-base-finetuned-cord-v1", "naver-clova-ix/donut-base-finetuned-cord-v1-2560", ]: _lowerCamelCase = '''<s_cord>''' elif model_name == "naver-clova-ix/donut-base-finetuned-cord-v2": _lowerCamelCase = '''s_cord-v2>''' elif model_name == "naver-clova-ix/donut-base-finetuned-zhtrainticket": _lowerCamelCase = '''<s_zhtrainticket>''' elif model_name in ["naver-clova-ix/donut-proto", "naver-clova-ix/donut-base"]: # use a random prompt _lowerCamelCase = '''hello world''' else: raise ValueError('''Model name not supported''' ) _lowerCamelCase = original_model.decoder.tokenizer(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors='''pt''' )[ '''input_ids''' ] _lowerCamelCase = original_model.encoder.model.patch_embed(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = model.encoder.embeddings(__UpperCAmelCase ) assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) # verify encoder hidden states _lowerCamelCase = original_model.encoder(__UpperCAmelCase ) _lowerCamelCase = model.encoder(__UpperCAmelCase ).last_hidden_state assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) # verify decoder hidden states _lowerCamelCase = original_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ).logits _lowerCamelCase = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ).logits assert torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: model.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) processor.push_to_hub('''nielsr/''' + model_name.split('''/''' )[-1] , commit_message='''Update model''' ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='naver-clova-ix/donut-base-finetuned-docvqa', required=False, type=str, help='Name of the original model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, required=False, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub.', ) snake_case__ = parser.parse_args() convert_donut_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
638
1
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['input_features'] def __init__( self , A_=80 , A_=1_60_00 , A_=1_60 , A_=30 , A_=4_00 , A_=0.0 , A_=False , **A_ , ) -> Union[str, Any]: """simple docstring""" super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , return_attention_mask=A_ , **A_ , ) _lowerCamelCase = n_fft _lowerCamelCase = hop_length _lowerCamelCase = chunk_length _lowerCamelCase = chunk_length * sampling_rate _lowerCamelCase = self.n_samples // hop_length _lowerCamelCase = sampling_rate _lowerCamelCase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=A_ , norm='''slaney''' , mel_scale='''slaney''' , ) def UpperCamelCase_ ( self , A_ ) -> np.ndarray: """simple docstring""" _lowerCamelCase = spectrogram( A_ , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) _lowerCamelCase = log_spec[:, :-1] _lowerCamelCase = np.maximum(A_ , log_spec.max() - 8.0 ) _lowerCamelCase = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCamelCase_ ( A_ , A_ , A_ = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: _lowerCamelCase = np.array(A_ , np.intaa ) _lowerCamelCase = [] for vector, length in zip(A_ , attention_mask.sum(-1 ) ): _lowerCamelCase = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _lowerCamelCase = padding_value normed_input_values.append(A_ ) else: _lowerCamelCase = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self , A_ , A_ = True , A_ = None , A_ = None , A_ = None , A_ = "max_length" , A_ = None , A_ = None , A_ = None , **A_ , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) _lowerCamelCase = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) _lowerCamelCase = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _lowerCamelCase = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): _lowerCamelCase = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _lowerCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _lowerCamelCase = [np.asarray([raw_speech] ).T] _lowerCamelCase = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding _lowerCamelCase = self.pad( A_ , padding=A_ , max_length=max_length if max_length else self.n_samples , truncation=A_ , pad_to_multiple_of=A_ , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: _lowerCamelCase = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) _lowerCamelCase = np.stack(padded_inputs['''input_features'''] , axis=0 ) # make sure list is in array format _lowerCamelCase = padded_inputs.get('''input_features''' ).transpose(2 , 0 , 1 ) _lowerCamelCase = [self._np_extract_fbank_features(A_ ) for waveform in input_features[0]] if isinstance(input_features[0] , A_ ): _lowerCamelCase = [np.asarray(A_ , dtype=np.floataa ) for feature in input_features] else: _lowerCamelCase = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _lowerCamelCase = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: _lowerCamelCase = padded_inputs.convert_to_tensors(A_ ) return padded_inputs def UpperCamelCase_ ( self ) -> Dict[str, Any]: """simple docstring""" _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
638
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
638
1
from __future__ import annotations import queue class UpperCamelCase : '''simple docstring''' def __init__( self , A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = data _lowerCamelCase = None _lowerCamelCase = None def __magic_name__( ) -> TreeNode: '''simple docstring''' print('''\n********Press N to stop entering at any point of time********\n''' ) _lowerCamelCase = input('''Enter the value of the root node: ''' ).strip().lower() _lowerCamelCase = queue.Queue() _lowerCamelCase = TreeNode(int(__UpperCAmelCase ) ) q.put(__UpperCAmelCase ) while not q.empty(): _lowerCamelCase = q.get() _lowerCamelCase = F'Enter the left node of {node_found.data}: ' _lowerCamelCase = input(__UpperCAmelCase ).strip().lower() or '''n''' if check == "n": return tree_node _lowerCamelCase = TreeNode(int(__UpperCAmelCase ) ) _lowerCamelCase = left_node q.put(__UpperCAmelCase ) _lowerCamelCase = F'Enter the right node of {node_found.data}: ' _lowerCamelCase = input(__UpperCAmelCase ).strip().lower() or '''n''' if check == "n": return tree_node _lowerCamelCase = TreeNode(int(__UpperCAmelCase ) ) _lowerCamelCase = right_node q.put(__UpperCAmelCase ) raise def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return _lowerCamelCase = queue.Queue() q.put(__UpperCAmelCase ) while not q.empty(): _lowerCamelCase = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return _lowerCamelCase = queue.Queue() q.put(__UpperCAmelCase ) while not q.empty(): _lowerCamelCase = [] while not q.empty(): _lowerCamelCase = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__UpperCAmelCase ) def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return _lowerCamelCase = [] _lowerCamelCase = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(__UpperCAmelCase ) _lowerCamelCase = n.left # end of while means current node doesn't have left child _lowerCamelCase = stack.pop() # start to traverse its right child _lowerCamelCase = n.right def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return _lowerCamelCase = [] _lowerCamelCase = node while n or stack: while n: stack.append(__UpperCAmelCase ) _lowerCamelCase = n.left _lowerCamelCase = stack.pop() print(n.data , end=''',''' ) _lowerCamelCase = n.right def __magic_name__( __UpperCAmelCase ) -> None: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not node: return _lowerCamelCase , _lowerCamelCase = [], [] _lowerCamelCase = node stacka.append(__UpperCAmelCase ) while stacka: # to find the reversed order of post order, store it in stack2 _lowerCamelCase = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__UpperCAmelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def __magic_name__( __UpperCAmelCase = "" , __UpperCAmelCase=50 , __UpperCAmelCase="*" ) -> str: '''simple docstring''' if not s: return "\n" + width * char _lowerCamelCase , _lowerCamelCase = divmod(width - len(__UpperCAmelCase ) - 2 , 2 ) return F'{left * char} {s} {(left + extra) * char}' if __name__ == "__main__": import doctest doctest.testmod() print(prompt('Binary Tree Traversals')) snake_case__ = build_tree() print(prompt('Pre Order Traversal')) pre_order(node) print(prompt() + '\n') print(prompt('In Order Traversal')) in_order(node) print(prompt() + '\n') print(prompt('Post Order Traversal')) post_order(node) print(prompt() + '\n') print(prompt('Level Order Traversal')) level_order(node) print(prompt() + '\n') print(prompt('Actual Level Order Traversal')) level_order_actual(node) print('*' * 50 + '\n') print(prompt('Pre Order Traversal - Iteration Version')) pre_order_iter(node) print(prompt() + '\n') print(prompt('In Order Traversal - Iteration Version')) in_order_iter(node) print(prompt() + '\n') print(prompt('Post Order Traversal - Iteration Version')) post_order_iter(node) print(prompt())
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
638
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = 42 class UpperCamelCase ( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , A_ = 6_55_36 , A_ = None , A_ = 2 , A_ = 2 , A_ = 0 , A_ = "fourier" , A_ = True , A_ = False , A_ = 0.0 , A_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A_ = "UNetMidBlock1D" , A_ = None , A_ = (32, 32, 64) , A_ = None , A_ = 8 , A_ = 1 , A_ = False , ) -> Dict: """simple docstring""" super().__init__() _lowerCamelCase = sample_size # time if time_embedding_type == "fourier": _lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A_ , log=A_ , flip_sin_to_cos=A_ ) _lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": _lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A_ , downscale_freq_shift=A_ ) _lowerCamelCase = block_out_channels[0] if use_timestep_embedding: _lowerCamelCase = block_out_channels[0] * 4 _lowerCamelCase = TimestepEmbedding( in_channels=A_ , time_embed_dim=A_ , act_fn=A_ , out_dim=block_out_channels[0] , ) _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = None # down _lowerCamelCase = in_channels for i, down_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_down_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A_ ) # mid _lowerCamelCase = get_mid_block( A_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A_ , add_downsample=A_ , ) # up _lowerCamelCase = list(reversed(A_ ) ) _lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: _lowerCamelCase = out_channels else: _lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(A_ ): _lowerCamelCase = output_channel _lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(A_ ) - 1 else final_upsample_channels ) _lowerCamelCase = i == len(A_ ) - 1 _lowerCamelCase = get_up_block( A_ , num_layers=A_ , in_channels=A_ , out_channels=A_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A_ ) _lowerCamelCase = output_channel # out _lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _lowerCamelCase = get_out_block( out_block_type=A_ , num_groups_out=A_ , embed_dim=block_out_channels[0] , out_channels=A_ , act_fn=A_ , fc_dim=block_out_channels[-1] // 4 , ) def UpperCamelCase_ ( self , A_ , A_ , A_ = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" _lowerCamelCase = timestep if not torch.is_tensor(A_ ): _lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A_ ) and len(timesteps.shape ) == 0: _lowerCamelCase = timesteps[None].to(sample.device ) _lowerCamelCase = self.time_proj(A_ ) if self.config.use_timestep_embedding: _lowerCamelCase = self.time_mlp(A_ ) else: _lowerCamelCase = timestep_embed[..., None] _lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _lowerCamelCase = () for downsample_block in self.down_blocks: _lowerCamelCase , _lowerCamelCase = downsample_block(hidden_states=A_ , temb=A_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _lowerCamelCase = self.mid_block(A_ , A_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _lowerCamelCase = down_block_res_samples[-1:] _lowerCamelCase = down_block_res_samples[:-1] _lowerCamelCase = upsample_block(A_ , res_hidden_states_tuple=A_ , temb=A_ ) # 5. post-process if self.out_block: _lowerCamelCase = self.out_block(A_ , A_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A_ )
638
1
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging snake_case__ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] snake_case__ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) snake_case__ = ' Hello world! cécé herlolip' snake_case__ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', ] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' _lowerCamelCase = dct.pop(__UpperCAmelCase ) _lowerCamelCase = val def __magic_name__( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = torch.load(__UpperCAmelCase , map_location='''cpu''' ) _lowerCamelCase = torch.hub.load('''pytorch/fairseq''' , '''bart.large.cnn''' ).eval() hub_interface.model.load_state_dict(sd['''model'''] ) return hub_interface def __magic_name__( __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase , _lowerCamelCase = emb.weight.shape _lowerCamelCase = nn.Linear(__UpperCAmelCase , __UpperCAmelCase , bias=__UpperCAmelCase ) _lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ) -> Union[str, Any]: '''simple docstring''' if not os.path.exists(__UpperCAmelCase ): _lowerCamelCase = torch.hub.load('''pytorch/fairseq''' , __UpperCAmelCase ).eval() else: _lowerCamelCase = load_xsum_checkpoint(__UpperCAmelCase ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: _lowerCamelCase = checkpoint_path.replace('''.''' , '''-''' ) _lowerCamelCase = BartConfig.from_pretrained(__UpperCAmelCase ) _lowerCamelCase = bart.encode(__UpperCAmelCase ).unsqueeze(0 ) _lowerCamelCase = BartTokenizer.from_pretrained(__UpperCAmelCase ).encode(__UpperCAmelCase , return_tensors='''pt''' ).unsqueeze(0 ) if not torch.eq(__UpperCAmelCase , __UpperCAmelCase ).all(): raise ValueError( F'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": _lowerCamelCase = bart.state_dict() remove_ignore_keys_(__UpperCAmelCase ) _lowerCamelCase = state_dict['''model.decoder.embed_tokens.weight'''] for src, dest in mnli_rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _lowerCamelCase = BartForSequenceClassification(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) _lowerCamelCase = bart.predict('''mnli''' , __UpperCAmelCase , return_logits=__UpperCAmelCase ) _lowerCamelCase = model(__UpperCAmelCase )[0] # logits else: # no classification heads to worry about _lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(__UpperCAmelCase ) _lowerCamelCase = state_dict['''decoder.embed_tokens.weight'''] _lowerCamelCase = bart.extract_features(__UpperCAmelCase ) if hf_checkpoint_name == "facebook/bart-large": _lowerCamelCase = BartModel(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) _lowerCamelCase = model(__UpperCAmelCase ).model[0] else: _lowerCamelCase = BartForConditionalGeneration(__UpperCAmelCase ).eval() # an existing summarization ckpt model.model.load_state_dict(__UpperCAmelCase ) if hasattr(__UpperCAmelCase , '''lm_head''' ): _lowerCamelCase = make_linear_from_emb(model.model.shared ) _lowerCamelCase = model.model(__UpperCAmelCase )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( F'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError('''Some values in `fairseq_output` are different from `new_model_outputs`''' ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) snake_case__ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
638
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () snake_case__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). snake_case__ = [0, 25, 50] snake_case__ = [25, 50, 75] snake_case__ = fuzz.membership.trimf(X, abca) snake_case__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. snake_case__ = np.ones(75) snake_case__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) snake_case__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) snake_case__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) snake_case__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] snake_case__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) snake_case__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] snake_case__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
638
1
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel snake_case__ = { 'text_branch': 'text_model', 'audio_branch': 'audio_model.audio_encoder', 'attn': 'attention.self', 'self.proj': 'output.dense', 'attention.self_mask': 'attn_mask', 'mlp.fc1': 'intermediate.dense', 'mlp.fc2': 'output.dense', 'norm1': 'layernorm_before', 'norm2': 'layernorm_after', 'bn0': 'batch_norm', } snake_case__ = AutoFeatureExtractor.from_pretrained('laion/clap-htsat-unfused', truncation='rand_trunc') def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=False ) -> List[str]: '''simple docstring''' _lowerCamelCase , _lowerCamelCase = create_model( '''HTSAT-tiny''' , '''roberta''' , __UpperCAmelCase , precision='''fp32''' , device='''cuda:0''' if torch.cuda.is_available() else '''cpu''' , enable_fusion=__UpperCAmelCase , fusion_type='''aff_2d''' if enable_fusion else None , ) return model, model_cfg def __magic_name__( __UpperCAmelCase ) -> Any: '''simple docstring''' _lowerCamelCase = {} _lowerCamelCase = r'''.*sequential.(\d+).*''' _lowerCamelCase = r'''.*_projection.(\d+).*''' for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _lowerCamelCase = key.replace(__UpperCAmelCase , __UpperCAmelCase ) if re.match(__UpperCAmelCase , __UpperCAmelCase ): # replace sequential layers with list _lowerCamelCase = re.match(__UpperCAmelCase , __UpperCAmelCase ).group(1 ) _lowerCamelCase = key.replace(F'sequential.{sequential_layer}.' , F'layers.{int(__UpperCAmelCase )//3}.linear.' ) elif re.match(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = int(re.match(__UpperCAmelCase , __UpperCAmelCase ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... _lowerCamelCase = 1 if projecton_layer == 0 else 2 _lowerCamelCase = key.replace(F'_projection.{projecton_layer}.' , F'_projection.linear{transformers_projection_layer}.' ) if "audio" and "qkv" in key: # split qkv into query key and value _lowerCamelCase = value _lowerCamelCase = mixed_qkv.size(0 ) // 3 _lowerCamelCase = mixed_qkv[:qkv_dim] _lowerCamelCase = mixed_qkv[qkv_dim : qkv_dim * 2] _lowerCamelCase = mixed_qkv[qkv_dim * 2 :] _lowerCamelCase = query_layer _lowerCamelCase = key_layer _lowerCamelCase = value_layer else: _lowerCamelCase = value return model_state_dict def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' _lowerCamelCase , _lowerCamelCase = init_clap(__UpperCAmelCase , enable_fusion=__UpperCAmelCase ) clap_model.eval() _lowerCamelCase = clap_model.state_dict() _lowerCamelCase = rename_state_dict(__UpperCAmelCase ) _lowerCamelCase = ClapConfig() _lowerCamelCase = enable_fusion _lowerCamelCase = ClapModel(__UpperCAmelCase ) # ignore the spectrogram embedding layer model.load_state_dict(__UpperCAmelCase , strict=__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) transformers_config.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument('--enable_fusion', action='store_true', help='Whether to enable fusion or not') snake_case__ = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
638
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger() @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" _lowerCamelCase = len(list(m.modules() ) ) == 1 or isinstance(A_ , nn.Convad ) or isinstance(A_ , nn.BatchNormad ) if has_not_submodules: self.traced.append(A_ ) def __call__( self , A_ ) -> Tuple: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(A_ ) [x.remove() for x in self.handles] return self @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda A_ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class UpperCamelCase : '''simple docstring''' A_ = 42 A_ = 42 A_ = 0 A_ = field(default_factory=__lowercase ) A_ = field(default_factory=__lowercase ) def __call__( self , A_ ) -> List[Any]: """simple docstring""" _lowerCamelCase = Tracker(self.dest )(A_ ).parametrized _lowerCamelCase = Tracker(self.src )(A_ ).parametrized _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.src_skip , A_ ) ) _lowerCamelCase = list(filter(lambda A_ : type(A_ ) not in self.dest_skip , A_ ) ) if len(A_ ) != len(A_ ): raise Exception( F'Numbers of operations are different. Source module has {len(A_ )} operations while' F' destination module has {len(A_ )}.' ) for dest_m, src_m in zip(A_ , A_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): _lowerCamelCase = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ).eval() _lowerCamelCase = ResNetForImageClassification(__UpperCAmelCase ).eval() _lowerCamelCase = ModuleTransfer(src=__UpperCAmelCase , dest=__UpperCAmelCase ) _lowerCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCAmelCase ) assert torch.allclose(from_model(__UpperCAmelCase ) , our_model(__UpperCAmelCase ).logits ), "The model logits don't match the original one." _lowerCamelCase = F'resnet{"-".join(name.split("resnet" ) )}' print(__UpperCAmelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add model''' , use_temp_dir=__UpperCAmelCase , ) # we can use the convnext one _lowerCamelCase = AutoImageProcessor.from_pretrained('''facebook/convnext-base-224-22k-1k''' ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message='''Add image processor''' , use_temp_dir=__UpperCAmelCase , ) print(F'Pushed {checkpoint_name}' ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = True ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 1000 _lowerCamelCase = (1, num_labels) _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = num_labels _lowerCamelCase = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = partial(__UpperCAmelCase , num_labels=__UpperCAmelCase , idalabel=__UpperCAmelCase , labelaid=__UpperCAmelCase ) _lowerCamelCase = { '''resnet18''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet26''': ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet34''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type='''basic''' ), '''resnet50''': ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet101''': ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), '''resnet152''': ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type='''bottleneck''' ), } if model_name: convert_weight_and_push(__UpperCAmelCase , names_to_config[model_name] , __UpperCAmelCase , __UpperCAmelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, expected_shape if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) snake_case__ = parser.parse_args() snake_case__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
638
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() _lowerCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) _lowerCamelCase = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } _lowerCamelCase = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub _lowerCamelCase = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase_ ( self , **A_ ) -> str: """simple docstring""" _lowerCamelCase = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> Optional[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase_ ( self , **A_ ) -> int: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = feature_extractor(A_ , return_tensors='''np''' ) _lowerCamelCase = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = '''This is a test string''' _lowerCamelCase = processor(text=A_ ) _lowerCamelCase = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase_ ( self , A_=(2, 10, 16) , A_=77 ) -> Optional[Any]: """simple docstring""" np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits(shape=(10, 16) , seed=13 ) _lowerCamelCase = processor.decode(A_ ) _lowerCamelCase = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: _lowerCamelCase = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: _lowerCamelCase = processor.batch_decode(A_ , A_ ) _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as p: _lowerCamelCase = decoder.decode_beams_batch(A_ , A_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 15 _lowerCamelCase = -20.0 _lowerCamelCase = -4.0 _lowerCamelCase = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] _lowerCamelCase = [d[0][2] for d in decoded_decoder_out] _lowerCamelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , A_ , atol=1E-3 ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = 2.0 _lowerCamelCase = 5.0 _lowerCamelCase = -20.0 _lowerCamelCase = True _lowerCamelCase = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) _lowerCamelCase = decoded_processor_out.text _lowerCamelCase = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: _lowerCamelCase = decoder.decode_beams_batch( A_ , A_ , ) _lowerCamelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = snapshot_download('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained(A_ ) _lowerCamelCase = processor.decoder.model_container[processor.decoder._model_key] _lowerCamelCase = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() _lowerCamelCase = os.listdir(A_ ) _lowerCamelCase = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = floats_list((3, 10_00) ) _lowerCamelCase = processor_wavaveca(A_ , return_tensors='''np''' ) _lowerCamelCase = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor_wavaveca.batch_decode(A_ ) _lowerCamelCase = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = self.get_feature_extractor() _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_decoder() _lowerCamelCase = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase_ ( A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = [d[key] for d in offsets] return retrieved_list def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits()[0] _lowerCamelCase = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" _lowerCamelCase = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) _lowerCamelCase = self._get_dummy_logits() _lowerCamelCase = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" import torch _lowerCamelCase = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) _lowerCamelCase = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) _lowerCamelCase = iter(A_ ) _lowerCamelCase = next(A_ ) _lowerCamelCase = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) _lowerCamelCase = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train _lowerCamelCase = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): _lowerCamelCase = model(A_ ).logits.cpu().numpy() _lowerCamelCase = processor.decode(logits[0] , output_word_offsets=A_ ) _lowerCamelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate _lowerCamelCase = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] _lowerCamelCase = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) _lowerCamelCase = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off _lowerCamelCase = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) _lowerCamelCase = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
638
1
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' _lowerCamelCase = len(__UpperCAmelCase ) _lowerCamelCase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _lowerCamelCase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _lowerCamelCase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _lowerCamelCase = subset[i - 1][j] if arr[i - 1] <= j: _lowerCamelCase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
638
1
import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def __magic_name__( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' def wrapper(*__UpperCAmelCase , **__UpperCAmelCase ): _lowerCamelCase = timeit.default_timer() _lowerCamelCase = func(*__UpperCAmelCase , **__UpperCAmelCase ) _lowerCamelCase = timeit.default_timer() - starttime return delta _lowerCamelCase = func.__name__ return wrapper def __magic_name__( __UpperCAmelCase , __UpperCAmelCase=100 , __UpperCAmelCase=None ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase = [] _lowerCamelCase = seq_shapes or {} for i in range(__UpperCAmelCase ): _lowerCamelCase = {} for col_id, (k, v) in enumerate(features.items() ): if isinstance(__UpperCAmelCase , _ArrayXD ): _lowerCamelCase = np.random.rand(*v.shape ).astype(v.dtype ) elif isinstance(__UpperCAmelCase , datasets.Value ): if v.dtype == "string": _lowerCamelCase = '''The small grey turtle was surprisingly fast when challenged.''' else: _lowerCamelCase = np.random.randint(10 , size=1 ).astype(v.dtype ).item() elif isinstance(__UpperCAmelCase , datasets.Sequence ): while isinstance(__UpperCAmelCase , datasets.Sequence ): _lowerCamelCase = v.feature _lowerCamelCase = seq_shapes[k] _lowerCamelCase = np.random.rand(*__UpperCAmelCase ).astype(v.dtype ) _lowerCamelCase = data dummy_data.append((i, example) ) return dummy_data def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=100 , __UpperCAmelCase=None ) -> Optional[int]: '''simple docstring''' _lowerCamelCase = generate_examples(__UpperCAmelCase , num_examples=__UpperCAmelCase , seq_shapes=__UpperCAmelCase ) with ArrowWriter(features=__UpperCAmelCase , path=__UpperCAmelCase ) as writer: for key, record in dummy_data: _lowerCamelCase = features.encode_example(__UpperCAmelCase ) writer.write(__UpperCAmelCase ) _lowerCamelCase , _lowerCamelCase = writer.finalize() if not num_final_examples == num_examples: raise ValueError( F'Error writing the dataset, wrote {num_final_examples} examples but should have written {num_examples}.' ) _lowerCamelCase = datasets.Dataset.from_file(filename=__UpperCAmelCase , info=datasets.DatasetInfo(features=__UpperCAmelCase ) ) return dataset
638
from typing import List import numpy as np def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' _lowerCamelCase = {key: len(__UpperCAmelCase ) for key, value in gen_kwargs.items() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _lowerCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , __UpperCAmelCase ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[range]: '''simple docstring''' _lowerCamelCase = [] for group_idx in range(__UpperCAmelCase ): _lowerCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _lowerCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _lowerCamelCase = range(__UpperCAmelCase , start + num_shards_to_add ) shards_indices_per_group.append(__UpperCAmelCase ) return shards_indices_per_group def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> List[dict]: '''simple docstring''' _lowerCamelCase = _number_of_shards_in_gen_kwargs(__UpperCAmelCase ) if num_shards == 1: return [dict(__UpperCAmelCase )] else: _lowerCamelCase = _distribute_shards(num_shards=__UpperCAmelCase , max_num_jobs=__UpperCAmelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(__UpperCAmelCase ) ) ] def __magic_name__( __UpperCAmelCase ) -> dict: '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , __UpperCAmelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> dict: '''simple docstring''' _lowerCamelCase = {len(__UpperCAmelCase ) for value in gen_kwargs.values() if isinstance(__UpperCAmelCase , __UpperCAmelCase )} _lowerCamelCase = {} for size in list_sizes: _lowerCamelCase = list(range(__UpperCAmelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _lowerCamelCase = dict(__UpperCAmelCase ) for key, value in shuffled_kwargs.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowerCamelCase = [value[i] for i in indices_per_size[len(__UpperCAmelCase )]] return shuffled_kwargs
638
1
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder snake_case__ = datasets.utils.logging.get_logger(__name__) class UpperCamelCase ( folder_based_builder.FolderBasedBuilderConfig ): '''simple docstring''' A_ = None A_ = None class UpperCamelCase ( folder_based_builder.FolderBasedBuilder ): '''simple docstring''' A_ = datasets.Audio() A_ = 'audio' A_ = AudioFolderConfig A_ = 42 # definition at the bottom of the script A_ = AudioClassification(audio_column='audio' , label_column='label' ) snake_case__ = [ '.aiff', '.au', '.avr', '.caf', '.flac', '.htk', '.svx', '.mat4', '.mat5', '.mpc2k', '.ogg', '.paf', '.pvf', '.raw', '.rf64', '.sd2', '.sds', '.ircam', '.voc', '.w64', '.wav', '.nist', '.wavex', '.wve', '.xi', '.mp3', '.opus', ] snake_case__ = AUDIO_EXTENSIONS
638
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=4_00 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 2_55 , A_=True , ) -> List[Any]: """simple docstring""" # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase_ ( self , A_ , A_=False ) -> List[str]: """simple docstring""" if not batched: _lowerCamelCase = image_inputs[0] if isinstance(A_ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(A_ , key=lambda A_ : item[0] )[0] _lowerCamelCase = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = YolosImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" _lowerCamelCase = YolosImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A_ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(A_ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" # Initialize image_processings _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = self.image_processing_class(do_resize=A_ , do_normalize=A_ , do_rescale=A_ ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _lowerCamelCase = image_processing_a.pad(A_ , return_tensors='''pt''' ) _lowerCamelCase = image_processing_a(A_ , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1E-4 ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them _lowerCamelCase = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = YolosImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area _lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) _lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1E-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels _lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks _lowerCamelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size _lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size _lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
638
1
from __future__ import annotations def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: # noqa: E741 '''simple docstring''' while r - l > 1: _lowerCamelCase = (l + r) // 2 if v[m] >= key: _lowerCamelCase = m else: _lowerCamelCase = m # noqa: E741 return r def __magic_name__( __UpperCAmelCase ) -> int: '''simple docstring''' if len(__UpperCAmelCase ) == 0: return 0 _lowerCamelCase = [0] * len(__UpperCAmelCase ) _lowerCamelCase = 1 _lowerCamelCase = v[0] for i in range(1 , len(__UpperCAmelCase ) ): if v[i] < tail[0]: _lowerCamelCase = v[i] elif v[i] > tail[length - 1]: _lowerCamelCase = v[i] length += 1 else: _lowerCamelCase = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
638
import argparse import json from tqdm import tqdm def __magic_name__( ) -> List[str]: '''simple docstring''' _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=__UpperCAmelCase , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=__UpperCAmelCase , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=__UpperCAmelCase , help='''where to store parsed gold_data_path file''' , ) _lowerCamelCase = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: _lowerCamelCase = json.load(__UpperCAmelCase ) for dpr_record in tqdm(__UpperCAmelCase ): _lowerCamelCase = dpr_record['''question'''] _lowerCamelCase = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(__UpperCAmelCase ) + '''\n''' ) if __name__ == "__main__": main()
638
1
from collections.abc import Sequence def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> float: '''simple docstring''' return sum(c * (x**i) for i, c in enumerate(__UpperCAmelCase ) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> float: '''simple docstring''' _lowerCamelCase = 0.0 for coeff in reversed(__UpperCAmelCase ): _lowerCamelCase = result * x + coeff return result if __name__ == "__main__": snake_case__ = (0.0, 0.0, 5.0, 9.3, 7.0) snake_case__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
638
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCamelCase : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ) -> None: """simple docstring""" _lowerCamelCase = device _lowerCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _lowerCamelCase = [0.48145466, 0.4578275, 0.40821073] _lowerCamelCase = [0.26862954, 0.26130258, 0.27577711] _lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCamelCase = torchvision.transforms.Resize(2_24 ) _lowerCamelCase = torchvision.transforms.CenterCrop(2_24 ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" _lowerCamelCase = self.resize(A_ ) _lowerCamelCase = self.center_crop(A_ ) _lowerCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ) -> Optional[Any]: """simple docstring""" _lowerCamelCase = self.tokenizer(text=A_ , **A_ ) _lowerCamelCase = self.preprocess_img(A_ ) _lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ) -> None: """simple docstring""" super().__init__() _lowerCamelCase = None _lowerCamelCase = device if device else get_device() if vqgan: _lowerCamelCase = vqgan else: _lowerCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _lowerCamelCase = clip else: _lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _lowerCamelCase = ProcessorGradientFlow(device=self.device ) _lowerCamelCase = iterations _lowerCamelCase = lr _lowerCamelCase = log _lowerCamelCase = make_grid _lowerCamelCase = return_val _lowerCamelCase = quantize _lowerCamelCase = self.vqgan.decoder.z_shape def UpperCamelCase_ ( self , A_=None , A_=None , A_=5 , A_=True ) -> Any: """simple docstring""" _lowerCamelCase = [] if output_path is None: _lowerCamelCase = '''./animation.gif''' if input_path is None: _lowerCamelCase = self.save_path _lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(A_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(A_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _lowerCamelCase = total_duration / len(A_ ) _lowerCamelCase = [frame_duration] * len(A_ ) if extend_frames: _lowerCamelCase = 1.5 _lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F'gif saved to {output_path}' ) def UpperCamelCase_ ( self , A_=None , A_=None ) -> Union[str, Any]: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _lowerCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _lowerCamelCase = preprocess_vqgan(A_ ) _lowerCamelCase , *_lowerCamelCase = self.vqgan.encode(A_ ) return z def UpperCamelCase_ ( self , A_ ) -> Optional[int]: """simple docstring""" _lowerCamelCase = self.latent.detach().requires_grad_() _lowerCamelCase = base_latent + transform_vector if self.quantize: _lowerCamelCase , *_lowerCamelCase = self.vqgan.quantize(A_ ) else: _lowerCamelCase = trans_latent return self.vqgan.decode(A_ ) def UpperCamelCase_ ( self , A_ , A_ , A_=None ) -> Any: """simple docstring""" _lowerCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors='''pt''' , padding=A_ ) _lowerCamelCase = self.clip(**A_ ) _lowerCamelCase = clip_outputs.logits_per_image if weights is not None: _lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Dict: """simple docstring""" _lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , A_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , A_ , weights=neg_prompts['''weights'''] ) else: _lowerCamelCase = torch.tensor([1] , device=self.device ) _lowerCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> str: """simple docstring""" _lowerCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCamelCase = self._add_vector(A_ ) _lowerCamelCase = loop_post_process(A_ ) _lowerCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print('''CLIP loss''' , A_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase_ ( self , A_ , A_ , A_ ) -> Any: """simple docstring""" wandb.init(reinit=A_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _lowerCamelCase = Image.open(A_ ) _lowerCamelCase = image.resize((2_56, 2_56) ) wandb.log('''Original Image''' , wandb.Image(A_ ) ) def UpperCamelCase_ ( self , A_ ) -> int: """simple docstring""" if not prompts: return [] _lowerCamelCase = [] _lowerCamelCase = [] if isinstance(A_ , A_ ): _lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _lowerCamelCase = prompt[0] _lowerCamelCase = float(prompt[1] ) elif ":" in prompt: _lowerCamelCase , _lowerCamelCase = prompt.split(''':''' ) _lowerCamelCase = float(A_ ) else: _lowerCamelCase = prompt _lowerCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def UpperCamelCase_ ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ) -> str: """simple docstring""" if image_path: _lowerCamelCase = self._get_latent(A_ ) else: _lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCamelCase = self.process_prompts(A_ ) _lowerCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(A_ ) _lowerCamelCase = save_path _lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(A_ ) ) _lowerCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({'''Image''': wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
638
1
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCamelCase ( __lowercase , unittest.TestCase ): '''simple docstring''' A_ = DDIMPipeline A_ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS A_ = PipelineTesterMixin.required_optional_params - { 'num_images_per_prompt', 'latents', 'callback', 'callback_steps', } A_ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS A_ = False def UpperCamelCase_ ( self ) -> str: """simple docstring""" torch.manual_seed(0 ) _lowerCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) _lowerCamelCase = DDIMScheduler() _lowerCamelCase = {'''unet''': unet, '''scheduler''': scheduler} return components def UpperCamelCase_ ( self , A_ , A_=0 ) -> int: """simple docstring""" if str(A_ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(A_ ) else: _lowerCamelCase = torch.Generator(device=A_ ).manual_seed(A_ ) _lowerCamelCase = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def UpperCamelCase_ ( self ) -> str: """simple docstring""" _lowerCamelCase = '''cpu''' _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = self.pipeline_class(**A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _lowerCamelCase = self.get_dummy_inputs(A_ ) _lowerCamelCase = pipe(**A_ ).images _lowerCamelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) _lowerCamelCase = np.array( [1.0_0_0E0_0, 5.7_1_7E-0_1, 4.7_1_7E-0_1, 1.0_0_0E0_0, 0.0_0_0E0_0, 1.0_0_0E0_0, 3.0_0_0E-0_4, 0.0_0_0E0_0, 9.0_0_0E-0_4] ) _lowerCamelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(A_ , 1E-3 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase = '''google/ddpm-cifar10-32''' _lowerCamelCase = UNetaDModel.from_pretrained(A_ ) _lowerCamelCase = DDIMScheduler() _lowerCamelCase = DDIMPipeline(unet=A_ , scheduler=A_ ) ddim.to(A_ ) ddim.set_progress_bar_config(disable=A_ ) _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = ddim(generator=A_ , eta=0.0 , output_type='''numpy''' ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase = np.array([0.1723, 0.1617, 0.1600, 0.1626, 0.1497, 0.1513, 0.1505, 0.1442, 0.1453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" _lowerCamelCase = '''google/ddpm-ema-bedroom-256''' _lowerCamelCase = UNetaDModel.from_pretrained(A_ ) _lowerCamelCase = DDIMScheduler.from_pretrained(A_ ) _lowerCamelCase = DDIMPipeline(unet=A_ , scheduler=A_ ) ddpm.to(A_ ) ddpm.set_progress_bar_config(disable=A_ ) _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = ddpm(generator=A_ , output_type='''numpy''' ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) _lowerCamelCase = np.array([0.0060, 0.0201, 0.0344, 0.0024, 0.0018, 0.0002, 0.0022, 0.0000, 0.0069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
638
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case__ = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
638
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' if curr_ind == len(__UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCAmelCase ) ): if valid_connection(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): # Insert current vertex into path as next transition _lowerCamelCase = next_ver # Validate created path if util_hamilton_cycle(__UpperCAmelCase , __UpperCAmelCase , curr_ind + 1 ): return True # Backtrack _lowerCamelCase = -1 return False def __magic_name__( __UpperCAmelCase , __UpperCAmelCase = 0 ) -> list[int]: '''simple docstring''' _lowerCamelCase = [-1] * (len(__UpperCAmelCase ) + 1) # initialize start and end of path with starting index _lowerCamelCase = _lowerCamelCase = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCAmelCase , __UpperCAmelCase , 1 ) else []
638
import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='%(message)s') def __magic_name__( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size, 1) ) def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.mean(1 ) # Centralize the data of class i _lowerCamelCase = data - column_reshape(__UpperCAmelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(__UpperCAmelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' _lowerCamelCase = features.mean(1 ) _lowerCamelCase = np.nan for i in range(__UpperCAmelCase ): _lowerCamelCase = features[:, labels == i] _lowerCamelCase = data.shape[1] _lowerCamelCase = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) _lowerCamelCase = device_data * np.dot( column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase ) , (column_reshape(__UpperCAmelCase ) - column_reshape(__UpperCAmelCase )).T , ) return covariance_sum / features.shape[1] def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' if features.any(): _lowerCamelCase = features.mean(1 ) # Center the dataset _lowerCamelCase = features - np.reshape(__UpperCAmelCase , (data_mean.size, 1) ) _lowerCamelCase = np.dot(__UpperCAmelCase , centered_data.T ) / features.shape[1] _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(__UpperCAmelCase ) # Take all the columns in the reverse order (-1), and then takes only the first _lowerCamelCase = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space _lowerCamelCase = np.dot(filtered_eigenvectors.T , __UpperCAmelCase ) logging.info('''Principal Component Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' assert classes > dimensions # Check if features have been already loaded if features.any: _lowerCamelCase , _lowerCamelCase = eigh( covariance_between_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , covariance_within_classes(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , ) _lowerCamelCase = eigenvectors[:, ::-1][:, :dimensions] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = np.linalg.svd(__UpperCAmelCase ) _lowerCamelCase = svd_matrix[:, 0:dimensions] _lowerCamelCase = np.dot(filtered_svd_matrix.T , __UpperCAmelCase ) logging.info('''Linear Discriminant Analysis computed''' ) return projected_data else: logging.basicConfig(level=logging.ERROR , format='''%(message)s''' , force=__UpperCAmelCase ) logging.error('''Dataset empty''' ) raise AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) _lowerCamelCase = np.array([0, 0, 0, 1, 1] ) _lowerCamelCase = 2 _lowerCamelCase = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = linear_discriminant_analysis( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if isinstance(__UpperCAmelCase , np.ndarray ): raise AssertionError( '''Did not raise AssertionError for dimensions > classes''' ) assert error_info.type is AssertionError def __magic_name__( ) -> None: '''simple docstring''' _lowerCamelCase = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) _lowerCamelCase = 2 _lowerCamelCase = np.array([[6.9_2_8_2_0_3_2_3, 8.6_6_0_2_5_4_0_4, 1_0.3_9_2_3_0_4_8_5], [3.0, 3.0, 3.0]] ) with pytest.raises(__UpperCAmelCase ) as error_info: _lowerCamelCase = principal_component_analysis(__UpperCAmelCase , __UpperCAmelCase ) if not np.allclose(__UpperCAmelCase , __UpperCAmelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
638
1