code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor lowerCamelCase : int = logging.get_logger(__name__) class __lowerCAmelCase (lowercase_ ): '''simple docstring''' def __init__(self : Tuple , *UpperCamelCase : Optional[Any] , **UpperCamelCase : Optional[Any] ): '''simple docstring''' warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , UpperCamelCase , ) super().__init__(*UpperCamelCase , **UpperCamelCase )
2
'''simple docstring''' import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand lowerCamelCase : Optional[Any] = ( '4S 3H 2C 7S 5H', '9D 8H 2C 6S 7H', '2D 6D 9D TH 7D', 'TC 8C 2S JH 6C', 'JH 8S TH AH QH', 'TS KS 5S 9S AC', 'KD 6S 9D TH AD', 'KS 8D 4D 9S 4S', # pair '8C 4S KH JS 4D', # pair 'QH 8H KD JH 8S', # pair 'KC 4H KS 2H 8D', # pair 'KD 4S KC 3H 8S', # pair 'AH 8S AS KC JH', # pair '3H 4C 4H 3S 2H', # 2 pairs '5S 5D 2C KH KH', # 2 pairs '3C KH 5D 5S KH', # 2 pairs 'AS 3C KH AD KH', # 2 pairs '7C 7S 3S 7H 5S', # 3 of a kind '7C 7S KH 2H 7H', # 3 of a kind 'AC KH QH AH AS', # 3 of a kind '2H 4D 3C AS 5S', # straight (low ace) '3C 5C 4C 2C 6H', # straight '6S 8S 7S 5H 9H', # straight 'JS QS 9H TS KH', # straight 'QC KH TS JS AH', # straight (high ace) '8C 9C 5C 3C TC', # flush '3S 8S 9S 5S KS', # flush '4C 5C 9C 8C KC', # flush 'JH 8H AH KH QH', # flush '3D 2H 3H 2C 2D', # full house '2H 2C 3S 3H 3D', # full house 'KH KC 3S 3H 3D', # full house 'JC 6H JS JD JH', # 4 of a kind 'JC 7H JS JD JH', # 4 of a kind 'JC KH JS JD JH', # 4 of a kind '2S AS 4S 5S 3S', # straight flush (low ace) '2D 6D 3D 4D 5D', # straight flush '5C 6C 3C 7C 4C', # straight flush 'JH 9H TH KH QH', # straight flush 'JH AH TH KH QH', # royal flush (high ace straight flush) ) lowerCamelCase : Tuple = ( ('2H 3H 4H 5H 6H', 'KS AS TS QS JS', 'Loss'), ('2H 3H 4H 5H 6H', 'AS AD AC AH JD', 'Win'), ('AS AH 2H AD AC', 'JS JD JC JH 3D', 'Win'), ('2S AH 2H AS AC', 'JS JD JC JH AD', 'Loss'), ('2S AH 2H AS AC', '2H 3H 5H 6H 7H', 'Win'), ('AS 3S 4S 8S 2S', '2H 3H 5H 6H 7H', 'Win'), ('2H 3H 5H 6H 7H', '2S 3H 4H 5S 6C', 'Win'), ('2S 3H 4H 5S 6C', '3D 4C 5H 6H 2S', 'Tie'), ('2S 3H 4H 5S 6C', 'AH AC 5H 6H AS', 'Win'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H AS', 'Loss'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H 7S', 'Win'), ('6S AD 7H 4S AS', 'AH AC 5H 6H 7S', 'Loss'), ('2S AH 4H 5S KC', 'AH AC 5H 6H 7S', 'Loss'), ('2S 3H 6H 7S 9C', '7H 3C TH 6H 9S', 'Loss'), ('4S 5H 6H TS AC', '3S 5H 6H TS AC', 'Win'), ('2S AH 4H 5S 6C', 'AD 4C 5H 6H 2C', 'Tie'), ('AS AH 3H AD AC', 'AS AH 2H AD AC', 'Win'), ('AH AC 5H 5C QS', 'AH AC 5H 5C KS', 'Loss'), ('AH AC 5H 5C QS', 'KH KC 5H 5C QS', 'Win'), ('7C 7S KH 2H 7H', '3C 3S AH 2H 3H', 'Win'), ('3C 3S AH 2H 3H', '7C 7S KH 2H 7H', 'Loss'), ('6H 5H 4H 3H 2H', '5H 4H 3H 2H AH', 'Win'), ('5H 4H 3H 2H AH', '5H 4H 3H 2H AH', 'Tie'), ('5H 4H 3H 2H AH', '6H 5H 4H 3H 2H', 'Loss'), ('AH AD KS KC AC', 'AH KD KH AC KC', 'Win'), ('2H 4D 3C AS 5S', '2H 4D 3C 6S 5S', 'Loss'), ('2H 3S 3C 3H 2S', '3S 3C 2S 2H 2D', 'Win'), ('4D 6D 5D 2D JH', '3S 8S 3H TC KH', 'Loss'), ('4S 6C 8S 3S 7S', 'AD KS 2D 7D 7C', 'Loss'), ('6S 4C 7H 8C 3H', '5H JC AH 9D 9C', 'Loss'), ('9D 9H JH TC QH', '3C 2S JS 5C 7H', 'Win'), ('2H TC 8S AD 9S', '4H TS 7H 2C 5C', 'Win'), ('9D 3S 2C 7S 7C', 'JC TD 3C TC 9H', 'Loss'), ) lowerCamelCase : Dict = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', True), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', False), ('AS 3S 4S 8S 2S', True), ) lowerCamelCase : Any = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', False), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', True), ) lowerCamelCase : Tuple = ( ('2H 4D 3C AS 5S', True, [5, 4, 3, 2, 14]), ('2H 5D 3C AS 5S', False, [14, 5, 5, 3, 2]), ('JH QD KC AS TS', False, [14, 13, 12, 11, 10]), ('9D 3S 2C 7S 7C', False, [9, 7, 7, 3, 2]), ) lowerCamelCase : Optional[int] = ( ('JH AH TH KH QH', 0), ('JH 9H TH KH QH', 0), ('JC KH JS JD JH', 7), ('KH KC 3S 3H 3D', 6), ('8C 9C 5C 3C TC', 0), ('JS QS 9H TS KH', 0), ('7C 7S KH 2H 7H', 3), ('3C KH 5D 5S KH', 2), ('QH 8H KD JH 8S', 1), ('2D 6D 9D TH 7D', 0), ) lowerCamelCase : Dict = ( ('JH AH TH KH QH', 23), ('JH 9H TH KH QH', 22), ('JC KH JS JD JH', 21), ('KH KC 3S 3H 3D', 20), ('8C 9C 5C 3C TC', 19), ('JS QS 9H TS KH', 18), ('7C 7S KH 2H 7H', 17), ('3C KH 5D 5S KH', 16), ('QH 8H KD JH 8S', 15), ('2D 6D 9D TH 7D', 14), ) def _SCREAMING_SNAKE_CASE () -> Union[str, Any]: """simple docstring""" lowercase__ ,lowercase__ = randrange(len(A ) ), randrange(len(A ) ) lowercase__ = ['''Loss''', '''Tie''', '''Win'''][(play >= oppo) + (play > oppo)] lowercase__ ,lowercase__ = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def _SCREAMING_SNAKE_CASE (A = 100 ) -> str: """simple docstring""" return (generate_random_hand() for _ in range(A )) @pytest.mark.parametrize('''hand, expected''' , A ) def _SCREAMING_SNAKE_CASE (A , A ) -> List[str]: """simple docstring""" assert PokerHand(A )._is_flush() == expected @pytest.mark.parametrize('''hand, expected''' , A ) def _SCREAMING_SNAKE_CASE (A , A ) -> Union[str, Any]: """simple docstring""" assert PokerHand(A )._is_straight() == expected @pytest.mark.parametrize('''hand, expected, card_values''' , A ) def _SCREAMING_SNAKE_CASE (A , A , A ) -> Any: """simple docstring""" lowercase__ = PokerHand(A ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('''hand, expected''' , A ) def _SCREAMING_SNAKE_CASE (A , A ) -> Tuple: """simple docstring""" assert PokerHand(A )._is_same_kind() == expected @pytest.mark.parametrize('''hand, expected''' , A ) def _SCREAMING_SNAKE_CASE (A , A ) -> Optional[Any]: """simple docstring""" assert PokerHand(A )._hand_type == expected @pytest.mark.parametrize('''hand, other, expected''' , A ) def _SCREAMING_SNAKE_CASE (A , A , A ) -> Union[str, Any]: """simple docstring""" assert PokerHand(A ).compare_with(PokerHand(A ) ) == expected @pytest.mark.parametrize('''hand, other, expected''' , generate_random_hands() ) def _SCREAMING_SNAKE_CASE (A , A , A ) -> Optional[Any]: """simple docstring""" assert PokerHand(A ).compare_with(PokerHand(A ) ) == expected def _SCREAMING_SNAKE_CASE () -> Tuple: """simple docstring""" lowercase__ = [PokerHand(A ) for hand in SORTED_HANDS] lowercase__ = poker_hands.copy() shuffle(A ) lowercase__ = chain(sorted(A ) ) for index, hand in enumerate(A ): assert hand == poker_hands[index] def _SCREAMING_SNAKE_CASE () -> List[Any]: """simple docstring""" lowercase__ = [PokerHand('''2D AC 3H 4H 5S''' ), PokerHand('''2S 3H 4H 5S 6C''' )] pokerhands.sort(reverse=A ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def _SCREAMING_SNAKE_CASE () -> int: """simple docstring""" lowercase__ = PokerHand('''2C 4S AS 3D 5C''' ) lowercase__ = True lowercase__ = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def _SCREAMING_SNAKE_CASE () -> Union[str, Any]: """simple docstring""" lowercase__ = 0 lowercase__ = os.path.abspath(os.path.dirname(A ) ) lowercase__ = os.path.join(A , '''poker_hands.txt''' ) with open(A ) as file_hand: for line in file_hand: lowercase__ = line[:14].strip() lowercase__ = line[15:].strip() lowercase__ ,lowercase__ = PokerHand(A ), PokerHand(A ) lowercase__ = player.compare_with(A ) if output == "Win": answer += 1 assert answer == 376
2
1
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class __a ( nn.Module ): _a : int _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = True _a : bool = False _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def UpperCAmelCase__ ( self ) -> List[str]: """simple docstring""" _UpperCAmelCase = [] _UpperCAmelCase = [] for i in range(self.num_layers ): _UpperCAmelCase = self.in_channels if i == 0 else self.out_channels _UpperCAmelCase = FlaxResnetBlockaD( in_channels=_SCREAMING_SNAKE_CASE , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnets _UpperCAmelCase = attentions if self.add_downsample: _UpperCAmelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> Tuple: """simple docstring""" _UpperCAmelCase = () for resnet, attn in zip(self.resnets , self.attentions ): _UpperCAmelCase = resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = attn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) output_states += (hidden_states,) if self.add_downsample: _UpperCAmelCase = self.downsamplers_a(_SCREAMING_SNAKE_CASE ) output_states += (hidden_states,) return hidden_states, output_states class __a ( nn.Module ): _a : int _a : int _a : float = 0.0 _a : int = 1 _a : bool = True _a : jnp.dtype = jnp.floataa def UpperCAmelCase__ ( self ) -> str: """simple docstring""" _UpperCAmelCase = [] for i in range(self.num_layers ): _UpperCAmelCase = self.in_channels if i == 0 else self.out_channels _UpperCAmelCase = FlaxResnetBlockaD( in_channels=_SCREAMING_SNAKE_CASE , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnets if self.add_downsample: _UpperCAmelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> List[str]: """simple docstring""" _UpperCAmelCase = () for resnet in self.resnets: _UpperCAmelCase = resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) output_states += (hidden_states,) if self.add_downsample: _UpperCAmelCase = self.downsamplers_a(_SCREAMING_SNAKE_CASE ) output_states += (hidden_states,) return hidden_states, output_states class __a ( nn.Module ): _a : int _a : int _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = True _a : bool = False _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def UpperCAmelCase__ ( self ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = [] _UpperCAmelCase = [] for i in range(self.num_layers ): _UpperCAmelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _UpperCAmelCase = self.prev_output_channel if i == 0 else self.out_channels _UpperCAmelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnets _UpperCAmelCase = attentions if self.add_upsample: _UpperCAmelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> Optional[Any]: """simple docstring""" for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states _UpperCAmelCase = res_hidden_states_tuple[-1] _UpperCAmelCase = res_hidden_states_tuple[:-1] _UpperCAmelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _UpperCAmelCase = resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = attn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) if self.add_upsample: _UpperCAmelCase = self.upsamplers_a(_SCREAMING_SNAKE_CASE ) return hidden_states class __a ( nn.Module ): _a : int _a : int _a : int _a : float = 0.0 _a : int = 1 _a : bool = True _a : jnp.dtype = jnp.floataa def UpperCAmelCase__ ( self ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = [] for i in range(self.num_layers ): _UpperCAmelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _UpperCAmelCase = self.prev_output_channel if i == 0 else self.out_channels _UpperCAmelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnets if self.add_upsample: _UpperCAmelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> List[Any]: """simple docstring""" for resnet in self.resnets: # pop res hidden states _UpperCAmelCase = res_hidden_states_tuple[-1] _UpperCAmelCase = res_hidden_states_tuple[:-1] _UpperCAmelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _UpperCAmelCase = resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) if self.add_upsample: _UpperCAmelCase = self.upsamplers_a(_SCREAMING_SNAKE_CASE ) return hidden_states class __a ( nn.Module ): _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def UpperCAmelCase__ ( self ) -> Any: """simple docstring""" _UpperCAmelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] _UpperCAmelCase = [] for _ in range(self.num_layers ): _UpperCAmelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnets _UpperCAmelCase = attentions def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> int: """simple docstring""" _UpperCAmelCase = self.resnets[0](_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): _UpperCAmelCase = attn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , deterministic=_SCREAMING_SNAKE_CASE ) return hidden_states
185
import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class __a ( UpperCAmelCase ): _a : Optional[int] = 'MCTCTFeatureExtractor' _a : int = 'AutoTokenizer' def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = self.feature_extractor _UpperCAmelCase = False def __call__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if "raw_speech" in kwargs: warnings.warn('Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.' ) _UpperCAmelCase = kwargs.pop('raw_speech' ) else: _UpperCAmelCase = kwargs.pop('audio' , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = kwargs.pop('sampling_rate' , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = kwargs.pop('text' , _SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: _UpperCAmelCase = args[0] _UpperCAmelCase = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if audio is not None: _UpperCAmelCase = self.feature_extractor(_SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if text is not None: _UpperCAmelCase = self.tokenizer(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if text is None: return inputs elif audio is None: return encodings else: _UpperCAmelCase = encodings['input_ids'] return inputs def UpperCAmelCase__ ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" if self._in_target_context_manager: return self.current_processor.pad(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = kwargs.pop('input_features' , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = kwargs.pop('labels' , _SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: _UpperCAmelCase = args[0] _UpperCAmelCase = args[1:] if input_features is not None: _UpperCAmelCase = self.feature_extractor.pad(_SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if labels is not None: _UpperCAmelCase = self.tokenizer.pad(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if labels is None: return input_features elif input_features is None: return labels else: _UpperCAmelCase = labels['input_ids'] return input_features def UpperCAmelCase__ ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" return self.tokenizer.decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @contextmanager def UpperCAmelCase__ ( self ) -> Optional[Any]: """simple docstring""" warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your audio inputs, or in a separate call.' ) _UpperCAmelCase = True _UpperCAmelCase = self.tokenizer yield _UpperCAmelCase = self.feature_extractor _UpperCAmelCase = False
185
1
import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _A ( unittest.TestCase): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=100 , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=30 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=3 , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = parent SCREAMING_SNAKE_CASE_ : Optional[int] = vocab_size SCREAMING_SNAKE_CASE_ : Tuple = batch_size SCREAMING_SNAKE_CASE_ : List[Any] = image_size SCREAMING_SNAKE_CASE_ : int = patch_size SCREAMING_SNAKE_CASE_ : Dict = num_channels SCREAMING_SNAKE_CASE_ : Optional[Any] = is_training SCREAMING_SNAKE_CASE_ : Any = use_labels SCREAMING_SNAKE_CASE_ : str = hidden_size SCREAMING_SNAKE_CASE_ : int = num_hidden_layers SCREAMING_SNAKE_CASE_ : Dict = num_attention_heads SCREAMING_SNAKE_CASE_ : str = intermediate_size SCREAMING_SNAKE_CASE_ : str = hidden_act SCREAMING_SNAKE_CASE_ : Union[str, Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[int] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[Any] = type_sequence_label_size SCREAMING_SNAKE_CASE_ : Union[str, Any] = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE_ : Union[str, Any] = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE_ : List[Any] = num_patches + 1 def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ : Optional[int] = None if self.use_labels: SCREAMING_SNAKE_CASE_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ : Dict = BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, pixel_values, labels def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = FlaxBeitModel(config=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = FlaxBeitForMaskedImageModeling(config=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[Any] = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.type_sequence_label_size SCREAMING_SNAKE_CASE_ : Any = FlaxBeitForImageClassification(config=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[Any] = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE_ : Dict = 1 SCREAMING_SNAKE_CASE_ : Dict = FlaxBeitForImageClassification(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ : List[str] = model(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ) : int = config_and_inputs SCREAMING_SNAKE_CASE_ : List[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_flax class _A ( __magic_name__ , unittest.TestCase): SCREAMING_SNAKE_CASE : str = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxBeitModelTester(self ) SCREAMING_SNAKE_CASE_ : List[Any] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def UpperCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ : Dict = model_class(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Tuple = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ : Any = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ : Optional[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): SCREAMING_SNAKE_CASE_ : List[Any] = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[Any] = model_class(_SCREAMING_SNAKE_CASE ) @jax.jit def model_jitted(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): return model(pixel_values=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) with self.subTest('JIT Enabled' ): SCREAMING_SNAKE_CASE_ : Dict = model_jitted(**_SCREAMING_SNAKE_CASE ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): SCREAMING_SNAKE_CASE_ : List[Any] = model_jitted(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) ) for jitted_output, output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self ): """simple docstring""" for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE_ : Optional[int] = model_class_name.from_pretrained('microsoft/beit-base-patch16-224' ) SCREAMING_SNAKE_CASE_ : List[str] = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def A_ ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @require_flax class _A ( unittest.TestCase): @cached_property def UpperCAmelCase ( self ): """simple docstring""" return BeitImageProcessor.from_pretrained('microsoft/beit-base-patch16-224' ) if is_vision_available() else None @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxBeitForMaskedImageModeling.from_pretrained('microsoft/beit-base-patch16-224-pt22k' ) SCREAMING_SNAKE_CASE_ : List[str] = self.default_image_processor SCREAMING_SNAKE_CASE_ : Tuple = prepare_img() SCREAMING_SNAKE_CASE_ : Any = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='np' ).pixel_values # prepare bool_masked_pos SCREAMING_SNAKE_CASE_ : str = np.ones((1, 196) , dtype=_SCREAMING_SNAKE_CASE ) # forward pass SCREAMING_SNAKE_CASE_ : str = model(pixel_values=_SCREAMING_SNAKE_CASE , bool_masked_pos=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Optional[Any] = outputs.logits # verify the logits SCREAMING_SNAKE_CASE_ : Any = (1, 196, 8192) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Optional[int] = np.array( [[-3.2437, 0.5072, -13.9174], [-3.2456, 0.4948, -13.9401], [-3.2033, 0.5121, -13.8550]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-2 ) ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = FlaxBeitForImageClassification.from_pretrained('microsoft/beit-base-patch16-224' ) SCREAMING_SNAKE_CASE_ : Tuple = self.default_image_processor SCREAMING_SNAKE_CASE_ : Any = prepare_img() SCREAMING_SNAKE_CASE_ : Any = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='np' ) # forward pass SCREAMING_SNAKE_CASE_ : Optional[Any] = model(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = outputs.logits # verify the logits SCREAMING_SNAKE_CASE_ : Any = (1, 1000) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Any = np.array([-1.2385, -1.0987, -1.0108] ) self.assertTrue(np.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 281 self.assertEqual(logits.argmax(-1 ).item() , _SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = FlaxBeitForImageClassification.from_pretrained('microsoft/beit-large-patch16-224-pt22k-ft22k' ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.default_image_processor SCREAMING_SNAKE_CASE_ : Optional[Any] = prepare_img() SCREAMING_SNAKE_CASE_ : Any = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='np' ) # forward pass SCREAMING_SNAKE_CASE_ : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[Any] = outputs.logits # verify the logits SCREAMING_SNAKE_CASE_ : List[Any] = (1, 2_1841) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : str = np.array([1.6881, -0.2787, 0.5901] ) self.assertTrue(np.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = 2396 self.assertEqual(logits.argmax(-1 ).item() , _SCREAMING_SNAKE_CASE )
253
import os import numpy import onnx def A_ ( a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = a.name SCREAMING_SNAKE_CASE_ : Dict = b.name SCREAMING_SNAKE_CASE_ : Optional[int] = '' SCREAMING_SNAKE_CASE_ : int = '' SCREAMING_SNAKE_CASE_ : Tuple = a == b SCREAMING_SNAKE_CASE_ : Dict = name_a SCREAMING_SNAKE_CASE_ : List[Any] = name_b return res def A_ ( a , a , a ): """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(a , a ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , a , a ) _graph_replace_input_with(node_proto.attribute[1].g , a , a ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , a , a ) def A_ ( a , a , a ): """simple docstring""" for n in graph_proto.node: _node_replace_input_with(a , a , a ) def A_ ( a , a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = list(model.graph.initializer ) SCREAMING_SNAKE_CASE_ : int = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i SCREAMING_SNAKE_CASE_ : List[Any] = inits[i].name SCREAMING_SNAKE_CASE_ : int = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , a , a ) def A_ ( a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = os.path.dirname(a ) SCREAMING_SNAKE_CASE_ : Tuple = os.path.basename(a ) SCREAMING_SNAKE_CASE_ : str = onnx.load(os.path.join(a , a ) ) SCREAMING_SNAKE_CASE_ : Dict = list(model.graph.initializer ) SCREAMING_SNAKE_CASE_ : str = set() SCREAMING_SNAKE_CASE_ : Optional[Any] = {} SCREAMING_SNAKE_CASE_ : Dict = [] SCREAMING_SNAKE_CASE_ : Dict = 0 for i in range(len(a ) ): if i in dup_set: continue for j in range(i + 1 , len(a ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(a ) dup_set.add(a ) SCREAMING_SNAKE_CASE_ : Optional[int] = inits[j].data_type SCREAMING_SNAKE_CASE_ : List[Any] = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 1_1: mem_size *= 8 else: print('unexpected data type: ' , a ) total_reduced_size += mem_size SCREAMING_SNAKE_CASE_ : Any = inits[i].name SCREAMING_SNAKE_CASE_ : Tuple = inits[j].name if name_i in dup_map: dup_map[name_i].append(a ) else: SCREAMING_SNAKE_CASE_ : Tuple = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1_0_2_4 / 1_0_2_4 / 1_0_2_4 , 'GB' ) SCREAMING_SNAKE_CASE_ : Tuple = sorted(a ) _remove_dup_initializers_from_model(a , a , a ) SCREAMING_SNAKE_CASE_ : List[Any] = 'optimized_' + model_file_name SCREAMING_SNAKE_CASE_ : Any = os.path.join(a , a ) onnx.save(a , a ) return new_model
253
1
'''simple docstring''' import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin _lowercase = get_tests_dir("""fixtures/test_sentencepiece_bpe.model""") class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' _lowercase : Optional[Any] = BartphoTokenizer _lowercase : List[str] = False _lowercase : Optional[Any] = True def _lowercase ( self ): """simple docstring""" super().setUp() _lowerCAmelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] _lowerCAmelCase = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) _lowerCAmelCase = {"""unk_token""": """<unk>"""} _lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) _lowerCAmelCase = BartphoTokenizer(_lowercase , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self , **_lowercase ): """simple docstring""" kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **_lowercase ) def _lowercase ( self , _lowercase ): """simple docstring""" _lowerCAmelCase = """This is a là test""" _lowerCAmelCase = """This is a<unk><unk> test""" return input_text, output_text def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = BartphoTokenizer(_lowercase , self.monolingual_vocab_file , **self.special_tokens_map ) _lowerCAmelCase = """This is a là test""" _lowerCAmelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() _lowerCAmelCase = tokenizer.tokenize(_lowercase ) self.assertListEqual(_lowercase , _lowercase ) _lowerCAmelCase = tokens + [tokenizer.unk_token] _lowerCAmelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ) , _lowercase )
229
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { """YituTech/conv-bert-base""": """https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json""", """YituTech/conv-bert-medium-small""": ( """https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json""" ), """YituTech/conv-bert-small""": """https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json""", # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : List[Any] = '''convbert''' def __init__( self , _lowercase=30_522 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3_072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=512 , _lowercase=2 , _lowercase=0.02 , _lowercase=1e-12 , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase=768 , _lowercase=2 , _lowercase=9 , _lowercase=1 , _lowercase=None , **_lowercase , ): """simple docstring""" super().__init__( pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase , ) _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = type_vocab_size _lowerCAmelCase = initializer_range _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = embedding_size _lowerCAmelCase = head_ratio _lowerCAmelCase = conv_kernel_size _lowerCAmelCase = num_groups _lowerCAmelCase = classifier_dropout class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _lowercase ( self ): """simple docstring""" if self.task == "multiple-choice": _lowerCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _lowerCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
229
1
'''simple docstring''' import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = DownBlockaD # noqa F405 lowerCamelCase__ = """down""" def A ( self : Dict ) -> Union[str, Any]: UpperCAmelCase : Any = [-0.02_32, -0.98_69, 0.80_54, -0.06_37, -0.16_88, -1.42_64, 0.44_70, -1.33_94, 0.09_04] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ResnetDownsampleBlockaD # noqa F405 lowerCamelCase__ = """down""" def A ( self : Optional[Any] ) -> Optional[int]: UpperCAmelCase : Dict = [0.07_10, 0.24_10, -0.73_20, -1.07_57, -1.13_43, 0.35_40, -0.01_33, -0.25_76, 0.09_48] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnDownBlockaD # noqa F405 lowerCamelCase__ = """down""" def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = [0.06_36, 0.89_64, -0.62_34, -1.01_31, 0.08_44, 0.49_35, 0.34_37, 0.09_11, -0.29_57] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = CrossAttnDownBlockaD # noqa F405 lowerCamelCase__ = """down""" def A ( self : int ) -> List[Any]: UpperCAmelCase , UpperCAmelCase : Union[str, Any] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : Dict = 32 return init_dict, inputs_dict def A ( self : List[Any] ) -> int: UpperCAmelCase : str = [0.22_38, -0.73_96, -0.22_55, -0.38_29, 0.19_25, 1.16_65, 0.06_03, -0.72_95, 0.19_83] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = SimpleCrossAttnDownBlockaD # noqa F405 lowerCamelCase__ = """down""" @property def A ( self : Optional[int] ) -> str: return super().get_dummy_input(include_encoder_hidden_states=__snake_case ) def A ( self : Optional[Any] ) -> str: UpperCAmelCase , UpperCAmelCase : Optional[int] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : int = 32 return init_dict, inputs_dict @unittest.skipIf(torch_device == '''mps''' , '''MPS result is not consistent''' ) def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = [0.79_21, -0.09_92, -0.19_62, -0.76_95, -0.42_42, 0.78_04, 0.47_37, 0.27_65, 0.33_38] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = SkipDownBlockaD # noqa F405 lowerCamelCase__ = """down""" @property def A ( self : Any ) -> Optional[Any]: return super().get_dummy_input(include_skip_sample=__snake_case ) def A ( self : Any ) -> Optional[Any]: UpperCAmelCase : Optional[Any] = [-0.08_45, -0.20_87, -0.24_65, 0.09_71, 0.19_00, -0.04_84, 0.26_64, 0.41_79, 0.50_69] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnSkipDownBlockaD # noqa F405 lowerCamelCase__ = """down""" @property def A ( self : List[Any] ) -> List[str]: return super().get_dummy_input(include_skip_sample=__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Any = [0.55_39, 0.16_09, 0.49_24, 0.05_37, -0.19_95, 0.40_50, 0.09_79, -0.27_21, -0.06_42] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = DownEncoderBlockaD # noqa F405 lowerCamelCase__ = """down""" @property def A ( self : Tuple ) -> Union[str, Any]: return super().get_dummy_input(include_temb=__snake_case ) def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Any = { '''in_channels''': 32, '''out_channels''': 32, } UpperCAmelCase : Any = self.dummy_input return init_dict, inputs_dict def A ( self : Optional[int] ) -> Optional[int]: UpperCAmelCase : Dict = [1.11_02, 0.53_02, 0.48_72, -0.00_23, -0.80_42, 0.04_83, -0.34_89, -0.56_32, 0.76_26] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnDownEncoderBlockaD # noqa F405 lowerCamelCase__ = """down""" @property def A ( self : str ) -> Union[str, Any]: return super().get_dummy_input(include_temb=__snake_case ) def A ( self : Dict ) -> Dict: UpperCAmelCase : Optional[int] = { '''in_channels''': 32, '''out_channels''': 32, } UpperCAmelCase : List[Any] = self.dummy_input return init_dict, inputs_dict def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[Any] = [0.89_66, -0.14_86, 0.85_68, 0.81_41, -0.90_46, -0.13_42, -0.09_72, -0.74_17, 0.15_38] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = UNetMidBlockaD # noqa F405 lowerCamelCase__ = """mid""" def A ( self : int ) -> str: UpperCAmelCase : List[Any] = { '''in_channels''': 32, '''temb_channels''': 128, } UpperCAmelCase : Optional[Any] = self.dummy_input return init_dict, inputs_dict def A ( self : List[str] ) -> Any: UpperCAmelCase : Dict = [-0.10_62, 1.72_48, 0.34_94, 1.45_69, -0.09_10, -1.24_21, -0.99_84, 0.67_36, 1.00_28] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = UNetMidBlockaDCrossAttn # noqa F405 lowerCamelCase__ = """mid""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : Optional[int] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : Any = 32 return init_dict, inputs_dict def A ( self : Any ) -> List[Any]: UpperCAmelCase : int = [0.01_87, 2.42_20, 0.44_84, 1.12_03, -0.61_21, -1.51_22, -0.82_70, 0.78_51, 1.83_35] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = UNetMidBlockaDSimpleCrossAttn # noqa F405 lowerCamelCase__ = """mid""" @property def A ( self : Optional[int] ) -> List[Any]: return super().get_dummy_input(include_encoder_hidden_states=__snake_case ) def A ( self : List[Any] ) -> str: UpperCAmelCase , UpperCAmelCase : Optional[Any] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : Union[str, Any] = 32 return init_dict, inputs_dict def A ( self : int ) -> str: UpperCAmelCase : Optional[int] = [0.71_43, 1.99_74, 0.54_48, 1.39_77, 0.12_82, -1.12_37, -1.42_38, 0.55_30, 0.88_80] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = UpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Optional[Any] ) -> Optional[int]: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) def A ( self : Dict ) -> List[str]: UpperCAmelCase : Tuple = [-0.20_41, -0.41_65, -0.30_22, 0.00_41, -0.66_28, -0.70_53, 0.19_28, -0.03_25, 0.05_23] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ResnetUpsampleBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Union[str, Any] ) -> Optional[Any]: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) def A ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase : Optional[Any] = [0.22_87, 0.35_49, -0.13_46, 0.47_97, -0.17_15, -0.96_49, 0.73_05, -0.58_64, -0.62_44] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = CrossAttnUpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Optional[int] ) -> List[Any]: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) def A ( self : Union[str, Any] ) -> int: UpperCAmelCase , UpperCAmelCase : Optional[Any] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : List[str] = 32 return init_dict, inputs_dict def A ( self : Any ) -> Optional[Any]: UpperCAmelCase : List[str] = [-0.14_03, -0.35_15, -0.04_20, -0.14_25, 0.31_67, 0.50_94, -0.21_81, 0.59_31, 0.55_82] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = SimpleCrossAttnUpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Dict ) -> int: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case , include_encoder_hidden_states=__snake_case ) def A ( self : Tuple ) -> Tuple: UpperCAmelCase , UpperCAmelCase : List[str] = super().prepare_init_args_and_inputs_for_common() UpperCAmelCase : List[str] = 32 return init_dict, inputs_dict def A ( self : int ) -> int: UpperCAmelCase : Optional[Any] = [0.26_45, 0.14_80, 0.09_09, 0.80_44, -0.97_58, -0.90_83, 0.09_94, -1.14_53, -0.74_02] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnUpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Optional[int] ) -> Tuple: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) @unittest.skipIf(torch_device == '''mps''' , '''MPS result is not consistent''' ) def A ( self : str ) -> Tuple: UpperCAmelCase : Any = [0.09_79, 0.13_26, 0.00_21, 0.06_59, 0.22_49, 0.00_59, 0.11_32, 0.59_52, 0.10_33] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = SkipUpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Optional[Any] ) -> Optional[int]: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : Optional[int] = [-0.08_93, -0.12_34, -0.15_06, -0.03_32, 0.01_23, -0.02_11, 0.05_66, 0.01_43, 0.03_62] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnSkipUpBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : Dict ) -> List[Any]: return super().get_dummy_input(include_res_hidden_states_tuple=__snake_case ) def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : Optional[int] = [0.03_61, 0.06_17, 0.27_87, -0.03_50, 0.03_42, 0.34_21, -0.08_43, 0.09_13, 0.30_15] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = UpDecoderBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : str ) -> Optional[int]: return super().get_dummy_input(include_temb=__snake_case ) def A ( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Tuple = {'''in_channels''': 32, '''out_channels''': 32} UpperCAmelCase : Union[str, Any] = self.dummy_input return init_dict, inputs_dict def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : int = [0.44_04, 0.19_98, -0.98_86, -0.33_20, -0.31_28, -0.70_34, -0.69_55, -0.23_38, -0.31_37] super().test_output(__snake_case ) class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AttnUpDecoderBlockaD # noqa F405 lowerCamelCase__ = """up""" @property def A ( self : str ) -> Tuple: return super().get_dummy_input(include_temb=__snake_case ) def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : Tuple = {'''in_channels''': 32, '''out_channels''': 32} UpperCAmelCase : Tuple = self.dummy_input return init_dict, inputs_dict def A ( self : Union[str, Any] ) -> Any: UpperCAmelCase : Any = [0.67_38, 0.44_91, 0.10_55, 1.07_10, 0.73_16, 0.33_39, 0.33_52, 0.10_23, 0.35_68] super().test_output(__snake_case )
23
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__: Optional[int] = logging.get_logger(__name__) def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: UpperCAmelCase : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: UpperCAmelCase : Tuple = 1024 UpperCAmelCase : List[Any] = 4096 UpperCAmelCase : str = 24 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = [5, 11, 17, 23] UpperCAmelCase : List[Any] = [256, 512, 1024, 1024] UpperCAmelCase : Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase : Optional[Any] = 768 UpperCAmelCase : Tuple = [1, 1, 1, 0.5] UpperCAmelCase : int = [256, 512, 768, 768] UpperCAmelCase : Any = 150 UpperCAmelCase : Tuple = 16 UpperCAmelCase : Any = (1, 384, 384) UpperCAmelCase : Optional[Any] = False UpperCAmelCase : Tuple = '''project''' if "ade" in checkpoint_url: UpperCAmelCase : Any = True UpperCAmelCase : str = 768 UpperCAmelCase : Optional[int] = [1, 1, 1, 0.5] UpperCAmelCase : List[Any] = 150 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = '''huggingface/label-files''' UpperCAmelCase : Tuple = '''ade20k-id2label.json''' UpperCAmelCase : Any = json.load(open(cached_download(hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) UpperCAmelCase : Optional[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase : List[Any] = idalabel UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} UpperCAmelCase : Union[str, Any] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : List[str] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Tuple ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase : Tuple = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: UpperCAmelCase : Union[str, Any] = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: UpperCAmelCase : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: UpperCAmelCase : Tuple = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: UpperCAmelCase : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: UpperCAmelCase : str = name.replace('''proj''' , '''projection''' ) if "blocks" in name: UpperCAmelCase : Any = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase : Optional[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase : Dict = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: UpperCAmelCase : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: UpperCAmelCase : str = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: UpperCAmelCase : Dict = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: UpperCAmelCase : int = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: UpperCAmelCase : Tuple = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: UpperCAmelCase : int = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: UpperCAmelCase : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase : str = name.replace(f"""refinenet{layer_idx}""" , f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase : List[str] = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: UpperCAmelCase : Union[str, Any] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: UpperCAmelCase : Any = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: UpperCAmelCase : Optional[int] = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: UpperCAmelCase : Tuple = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase : Optional[Any] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase : List[Any] = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase : str = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase : List[str] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase : Tuple = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: UpperCAmelCase : Dict = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: UpperCAmelCase : Any = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: UpperCAmelCase : Optional[int] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: UpperCAmelCase : Union[str, Any] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: UpperCAmelCase : Optional[int] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: UpperCAmelCase : Optional[Any] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: UpperCAmelCase : Optional[int] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: UpperCAmelCase : List[str] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase : List[Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: UpperCAmelCase : List[Any] = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase : Tuple = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] ) -> Optional[Any]: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[int] = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Tuple = in_proj_weight[: config.hidden_size, :] UpperCAmelCase : int = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( ) -> List[str]: UpperCAmelCase : Optional[int] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase : Optional[int] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] ) -> Any: UpperCAmelCase , UpperCAmelCase : int = get_dpt_config(_lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase : List[Any] = torch.load(_lowerCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(_lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase : Any = state_dict.pop(_lowerCAmelCase ) UpperCAmelCase : List[Any] = val # read in qkv matrices read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model UpperCAmelCase : Optional[Any] = DPTForSemanticSegmentation(_lowerCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase : int = 480 if '''ade''' in checkpoint_url else 384 UpperCAmelCase : List[Any] = DPTImageProcessor(size=_lowerCAmelCase ) UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Optional[int] = image_processor(_lowerCAmelCase , return_tensors='''pt''' ) # forward pass UpperCAmelCase : Any = model(**_lowerCAmelCase ).logits if '''ade''' in checkpoint_url else model(**_lowerCAmelCase ).predicted_depth if show_prediction: UpperCAmelCase : Dict = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=_lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) UpperCamelCase__: Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
23
1
import colorsys from PIL import Image # type: ignore def __lowerCamelCase (UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] ): SCREAMING_SNAKE_CASE = x SCREAMING_SNAKE_CASE = y for step in range(_lowerCamelCase ): # noqa: B007 SCREAMING_SNAKE_CASE = a * a - b * b + x SCREAMING_SNAKE_CASE = 2 * a * b + y SCREAMING_SNAKE_CASE = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def __lowerCamelCase (UpperCAmelCase__ : Dict ): if distance == 1: return (0, 0, 0) else: return (2_5_5, 2_5_5, 2_5_5) def __lowerCamelCase (UpperCAmelCase__ : Optional[Any] ): if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_5_5 ) for i in colorsys.hsv_to_rgb(_lowerCamelCase , 1 , 1 ) ) def __lowerCamelCase (UpperCAmelCase__ : Dict = 8_0_0 , UpperCAmelCase__ : List[Any] = 6_0_0 , UpperCAmelCase__ : Any = -0.6 , UpperCAmelCase__ : Optional[int] = 0 , UpperCAmelCase__ : str = 3.2 , UpperCAmelCase__ : Optional[Any] = 5_0 , UpperCAmelCase__ : List[Any] = True , ): SCREAMING_SNAKE_CASE = Image.new("RGB" , (image_width, image_height) ) SCREAMING_SNAKE_CASE = img.load() # loop through the image-coordinates for image_x in range(_lowerCamelCase ): for image_y in range(_lowerCamelCase ): # determine the figure-coordinates based on the image-coordinates SCREAMING_SNAKE_CASE = figure_width / image_width * image_height SCREAMING_SNAKE_CASE = figure_center_x + (image_x / image_width - 0.5) * figure_width SCREAMING_SNAKE_CASE = figure_center_y + (image_y / image_height - 0.5) * figure_height SCREAMING_SNAKE_CASE = get_distance(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: SCREAMING_SNAKE_CASE = get_color_coded_rgb(_lowerCamelCase ) else: SCREAMING_SNAKE_CASE = get_black_and_white_rgb(_lowerCamelCase ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _lowerCamelCase : List[Any] = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
363
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCamelCase : Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} _lowerCamelCase : List[Any] = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] _lowerCamelCase : Optional[Any] = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } _lowerCamelCase : Any = {f"""funnel-transformer/{name}""": 5_12 for name in _model_names} _lowerCamelCase : Optional[Any] = {f"""funnel-transformer/{name}""": {'''do_lower_case''': True} for name in _model_names} class lowercase ( a ): lowercase__ : Optional[int] = VOCAB_FILES_NAMES lowercase__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION lowercase__ : Union[str, Any] = FunnelTokenizer lowercase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : int = 2 def __init__( self : str , _UpperCamelCase : str=None , _UpperCamelCase : str=None , _UpperCamelCase : Union[str, Any]=True , _UpperCamelCase : str="<unk>" , _UpperCamelCase : Optional[Any]="<sep>" , _UpperCamelCase : Optional[int]="<pad>" , _UpperCamelCase : int="<cls>" , _UpperCamelCase : Dict="<mask>" , _UpperCamelCase : Union[str, Any]="<s>" , _UpperCamelCase : Optional[int]="</s>" , _UpperCamelCase : Dict=True , _UpperCamelCase : List[Any]=True , _UpperCamelCase : Any=None , _UpperCamelCase : Dict="##" , **_UpperCamelCase : Dict , ) -> Optional[int]: '''simple docstring''' super().__init__( _UpperCamelCase , tokenizer_file=_UpperCamelCase , do_lower_case=_UpperCamelCase , unk_token=_UpperCamelCase , sep_token=_UpperCamelCase , pad_token=_UpperCamelCase , cls_token=_UpperCamelCase , mask_token=_UpperCamelCase , bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , clean_text=_UpperCamelCase , tokenize_chinese_chars=_UpperCamelCase , strip_accents=_UpperCamelCase , wordpieces_prefix=_UpperCamelCase , **_UpperCamelCase , ) SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , _UpperCamelCase ) != do_lower_case or normalizer_state.get("strip_accents" , _UpperCamelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , _UpperCamelCase ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE = getattr(_UpperCamelCase , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE = do_lower_case SCREAMING_SNAKE_CASE = strip_accents SCREAMING_SNAKE_CASE = tokenize_chinese_chars SCREAMING_SNAKE_CASE = normalizer_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = do_lower_case def __snake_case( self : Union[str, Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Dict=None ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __snake_case( self : int , _UpperCamelCase : List[int] , _UpperCamelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.sep_token_id] SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __snake_case( self : Optional[Any] , _UpperCamelCase : str , _UpperCamelCase : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = self._tokenizer.model.save(_UpperCamelCase , name=_UpperCamelCase ) return tuple(_UpperCamelCase )
206
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(snake_case__ ) == 0: raise ValueError('''Input list must be a non empty list''' ) if len(snake_case__ ) == 1: return True A : Any = series[1] - series[0] for index in range(len(snake_case__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(snake_case__ ) == 0: raise ValueError('''Input list must be a non empty list''' ) A : Optional[Any] = 0 for val in series: answer += val return answer / len(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod()
3
from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class __snake_case : pass
51
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'facebook/xlm-roberta-xl': 'https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json', 'facebook/xlm-roberta-xxl': 'https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "xlm-roberta-xl" def __init__(self , UpperCAmelCase=250880 , UpperCAmelCase=2560 , UpperCAmelCase=36 , UpperCAmelCase=32 , UpperCAmelCase=10240 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=514 , UpperCAmelCase=1 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-0_5 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=2 , UpperCAmelCase="absolute" , UpperCAmelCase=True , UpperCAmelCase=None , **UpperCAmelCase , ) -> Any: super().__init__(pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = classifier_dropout class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
270
'''simple docstring''' import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[str]: if dst_width < 0 or dst_height < 0: raise ValueError("""Destination width/height should be > 0""" ) _snake_case = img _snake_case = img.shape[1] _snake_case = img.shape[0] _snake_case = dst_width _snake_case = dst_height _snake_case = self.src_w / self.dst_w _snake_case = self.src_h / self.dst_h _snake_case = _snake_case = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 255 ) def lowercase (self ) -> List[Any]: for i in range(self.dst_h ): for j in range(self.dst_w ): _snake_case = self.img[self.get_y(UpperCAmelCase )][self.get_x(UpperCAmelCase )] def lowercase (self , UpperCAmelCase ) -> int: return int(self.ratio_x * x ) def lowercase (self , UpperCAmelCase ) -> int: return int(self.ratio_y * y ) if __name__ == "__main__": __lowerCAmelCase , __lowerCAmelCase = 800, 600 __lowerCAmelCase = imread('image_data/lena.jpg', 1) __lowerCAmelCase = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( f'''Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}''', n.output ) waitKey(0) destroyAllWindows()
270
1
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase_ : str ) -> str: __lowerCamelCase : str = len(UpperCAmelCase_ ) while cur > 1: # Find the maximum number in arr __lowerCamelCase : Dict = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __lowerCamelCase : Any = arr[mi::-1] + arr[mi + 1 : len(UpperCAmelCase_ )] # Reverse whole list __lowerCamelCase : Dict = arr[cur - 1 :: -1] + arr[cur : len(UpperCAmelCase_ )] cur -= 1 return arr if __name__ == "__main__": A__ : List[Any] = input("""Enter numbers separated by a comma:\n""").strip() A__ : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
185
'''simple docstring''' from math import isqrt def UpperCAmelCase__ ( UpperCAmelCase_ : int ) -> bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(UpperCAmelCase_ ) + 1 ) ) def UpperCAmelCase__ ( UpperCAmelCase_ : int = 10**6 ) -> int: __lowerCamelCase : Optional[Any] = 0 __lowerCamelCase : Optional[int] = 1 __lowerCamelCase : List[str] = 7 while prime_candidate < max_prime: primes_count += is_prime(UpperCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(f'''{solution() = }''')
185
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class lowerCamelCase : """simple docstring""" def __init__( self : Optional[int] , __magic_name__ : Any , __magic_name__ : Optional[Any]=2 , __magic_name__ : Dict=True , __magic_name__ : int=False , __magic_name__ : int=10 , __magic_name__ : int=3 , __magic_name__ : int=32 * 4 , __magic_name__ : Optional[int]=32 * 6 , __magic_name__ : Tuple=4 , __magic_name__ : str=32 , ) -> List[str]: SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_auxiliary_loss SCREAMING_SNAKE_CASE_ = num_queries SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = min_size SCREAMING_SNAKE_CASE_ = max_size SCREAMING_SNAKE_CASE_ = num_labels SCREAMING_SNAKE_CASE_ = mask_feature_size def __A ( self : Optional[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) SCREAMING_SNAKE_CASE_ = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_a ) SCREAMING_SNAKE_CASE_ = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_a ) > 0.5 ).float() SCREAMING_SNAKE_CASE_ = (torch.rand((self.batch_size, self.num_labels) , device=_a ) > 0.5).long() SCREAMING_SNAKE_CASE_ = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __A ( self : Any ) -> int: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def __A ( self : List[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def __A ( self : Dict , __magic_name__ : Optional[int] , __magic_name__ : int ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = output.encoder_hidden_states SCREAMING_SNAKE_CASE_ = output.pixel_decoder_hidden_states SCREAMING_SNAKE_CASE_ = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) , config.decoder_config.decoder_layers ) def __A ( self : Dict , __magic_name__ : int , __magic_name__ : Tuple , __magic_name__ : Any , __magic_name__ : Tuple=False ) -> Optional[int]: with torch.no_grad(): SCREAMING_SNAKE_CASE_ = MaskFormerModel(config=_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE_ = model(pixel_values=_a , pixel_mask=_a ) SCREAMING_SNAKE_CASE_ = model(_a , output_hidden_states=_a ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a , _a ) def __A ( self : Union[str, Any] , __magic_name__ : List[Any] , __magic_name__ : Optional[int] , __magic_name__ : Dict , __magic_name__ : Optional[int] , __magic_name__ : int ) -> Tuple: SCREAMING_SNAKE_CASE_ = MaskFormerForInstanceSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(__magic_name__ : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(pixel_values=_a , pixel_mask=_a ) SCREAMING_SNAKE_CASE_ = model(_a ) comm_check_on_output(_a ) SCREAMING_SNAKE_CASE_ = model( pixel_values=_a , pixel_mask=_a , mask_labels=_a , class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowerCamelCase (__lowercase , __lowercase , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCamelCase__ = ( {'''feature-extraction''': MaskFormerModel, '''image-segmentation''': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def __A ( self : List[Any] ) -> Tuple: SCREAMING_SNAKE_CASE_ = MaskFormerModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=_a , has_text_modality=_a ) def __A ( self : List[Any] ) -> str: self.config_tester.run_common_tests() def __A ( self : List[str] ) -> Any: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_a , **_a , output_hidden_states=_a ) def __A ( self : int ) -> Dict: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason="MaskFormer does not use inputs_embeds" ) def __A ( self : Any ) -> Dict: pass @unittest.skip(reason="MaskFormer does not have a get_input_embeddings method" ) def __A ( self : Union[str, Any] ) -> str: pass @unittest.skip(reason="MaskFormer is not a generative model" ) def __A ( self : Tuple ) -> Dict: pass @unittest.skip(reason="MaskFormer does not use token embeddings" ) def __A ( self : List[Any] ) -> Tuple: pass @require_torch_multi_gpu @unittest.skip( reason="MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def __A ( self : List[str] ) -> List[str]: pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __A ( self : Dict ) -> Dict: pass def __A ( self : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_a ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a ) @slow def __A ( self : str ) -> List[str]: for model_name in ["facebook/maskformer-swin-small-coco"]: SCREAMING_SNAKE_CASE_ = MaskFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __A ( self : str ) -> List[str]: SCREAMING_SNAKE_CASE_ = (self.model_tester.min_size,) * 2 SCREAMING_SNAKE_CASE_ = { "pixel_values": torch.randn((2, 3, *size) , device=_a ), "mask_labels": torch.randn((2, 10, *size) , device=_a ), "class_labels": torch.zeros(2 , 10 , device=_a ).long(), } SCREAMING_SNAKE_CASE_ = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(_a ) SCREAMING_SNAKE_CASE_ = model(**_a ) self.assertTrue(outputs.loss is not None ) def __A ( self : Tuple ) -> Dict: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_a , **_a , output_hidden_states=_a ) def __A ( self : Dict ) -> Any: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_a ).to(_a ) SCREAMING_SNAKE_CASE_ = model(**_a , output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __A ( self : Any ) -> Any: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss SCREAMING_SNAKE_CASE_ = self.all_model_classes[1] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ = model_class(_a ) model.to(_a ) model.train() SCREAMING_SNAKE_CASE_ = model(_a , mask_labels=_a , class_labels=_a ).loss loss.backward() def __A ( self : int ) -> Tuple: # only MaskFormerForInstanceSegmentation has the loss SCREAMING_SNAKE_CASE_ = self.all_model_classes[1] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = model_class(_a ) model.to(_a ) model.train() SCREAMING_SNAKE_CASE_ = model(_a , mask_labels=_a , class_labels=_a ) SCREAMING_SNAKE_CASE_ = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() SCREAMING_SNAKE_CASE_ = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't SCREAMING_SNAKE_CASE_ = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() SCREAMING_SNAKE_CASE_ = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) A : int = 1E-4 def a__ ( ): SCREAMING_SNAKE_CASE_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class lowerCamelCase (unittest.TestCase ): """simple docstring""" @cached_property def __A ( self : List[Any] ) -> Tuple: return ( MaskFormerImageProcessor.from_pretrained("facebook/maskformer-swin-small-coco" ) if is_vision_available() else None ) def __A ( self : Dict ) -> int: SCREAMING_SNAKE_CASE_ = MaskFormerModel.from_pretrained("facebook/maskformer-swin-small-coco" ).to(_a ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(_a , return_tensors="pt" ).to(_a ) SCREAMING_SNAKE_CASE_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a , (1, 3, 800, 1_088) ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_a ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _a , atol=_a ) ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _a , atol=_a ) ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _a , atol=_a ) ) def __A ( self : Tuple ) -> List[str]: SCREAMING_SNAKE_CASE_ = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(_a ) .eval() ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(_a , return_tensors="pt" ).to(_a ) SCREAMING_SNAKE_CASE_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a , (1, 3, 800, 1_088) ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_a ) # masks_queries_logits SCREAMING_SNAKE_CASE_ = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) SCREAMING_SNAKE_CASE_ = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] SCREAMING_SNAKE_CASE_ = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _a , atol=_a ) ) # class_queries_logits SCREAMING_SNAKE_CASE_ = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) SCREAMING_SNAKE_CASE_ = torch.tensor( [ [1.6512e00, -5.2572e00, -3.3519e00], [3.6169e-02, -5.9025e00, -2.9313e00], [1.0766e-04, -7.7630e00, -5.1263e00], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _a , atol=_a ) ) def __A ( self : str ) -> str: SCREAMING_SNAKE_CASE_ = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-resnet101-coco-stuff" ) .to(_a ) .eval() ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(_a , return_tensors="pt" ).to(_a ) SCREAMING_SNAKE_CASE_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a , (1, 3, 800, 1_088) ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_a ) # masks_queries_logits SCREAMING_SNAKE_CASE_ = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) SCREAMING_SNAKE_CASE_ = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7_711]] SCREAMING_SNAKE_CASE_ = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _a , atol=_a ) ) # class_queries_logits SCREAMING_SNAKE_CASE_ = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _a , atol=_a ) ) def __A ( self : Optional[Any] ) -> int: SCREAMING_SNAKE_CASE_ = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(_a ) .eval() ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) SCREAMING_SNAKE_CASE_ = inputs["pixel_values"].to(_a ) SCREAMING_SNAKE_CASE_ = [el.to(_a ) for el in inputs["mask_labels"]] SCREAMING_SNAKE_CASE_ = [el.to(_a ) for el in inputs["class_labels"]] with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_a ) self.assertTrue(outputs.loss is not None )
350
from __future__ import annotations import numpy as np def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = np.shape(__UpperCamelCase ) if rows != columns: SCREAMING_SNAKE_CASE_ = ( "'table' has to be of square shaped array but got a " F'''{rows}x{columns} array:\n{table}''' ) raise ValueError(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = np.zeros((rows, columns) ) SCREAMING_SNAKE_CASE_ = np.zeros((rows, columns) ) for i in range(__UpperCamelCase ): for j in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE_ = sum(lower[i][k] * upper[k][j] for k in range(__UpperCamelCase ) ) if upper[j][j] == 0: raise ArithmeticError("No LU decomposition exists" ) SCREAMING_SNAKE_CASE_ = (table[i][j] - total) / upper[j][j] SCREAMING_SNAKE_CASE_ = 1 for j in range(__UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = sum(lower[i][k] * upper[k][j] for k in range(__UpperCamelCase ) ) SCREAMING_SNAKE_CASE_ = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
305
0
'''simple docstring''' def UpperCamelCase_ ( snake_case_ : int , snake_case_ : int ) -> int: '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def UpperCamelCase_ ( ) -> None: '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
229
'''simple docstring''' import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase_ ( snake_case_ : Any ) -> Optional[Any]: '''simple docstring''' __lowerCAmelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """decoder.output_projection.weight""", """_float_tensor""", """encoder.embed_positions._float_tensor""", """decoder.embed_positions._float_tensor""", ] for k in ignore_keys: state_dict.pop(snake_case_ , snake_case_ ) def UpperCamelCase_ ( snake_case_ : Optional[Any] ) -> List[Any]: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase = emb.weight.shape __lowerCAmelCase = nn.Linear(snake_case_ , snake_case_ , bias=snake_case_ ) __lowerCAmelCase = emb.weight.data return lin_layer def UpperCamelCase_ ( snake_case_ : Any ) -> Any: '''simple docstring''' __lowerCAmelCase = torch.load(snake_case_ , map_location="""cpu""" ) __lowerCAmelCase = mam_aaa["""args"""] or mam_aaa["""cfg"""]["""model"""] __lowerCAmelCase = mam_aaa["""model"""] remove_ignore_keys_(snake_case_ ) __lowerCAmelCase = state_dict["""encoder.embed_tokens.weight"""].shape[0] __lowerCAmelCase = MaMaaaConfig( vocab_size=snake_case_ , max_position_embeddings=10_24 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="""relu""" , ) __lowerCAmelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCAmelCase = MaMaaaForConditionalGeneration(snake_case_ ) model.model.load_state_dict(snake_case_ , strict=snake_case_ ) __lowerCAmelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _A : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument('''fairseq_path''', type=str, help='''path to a model.pt on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _A : str = parser.parse_args() _A : Optional[int] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
229
1
_SCREAMING_SNAKE_CASE : Union[str, Any] = [ "DownloadConfig", "DownloadManager", "DownloadMode", "StreamingDownloadManager", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
213
def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" snake_case = len(UpperCamelCase_ ) snake_case = [[0] * n for i in range(UpperCamelCase_ )] for i in range(UpperCamelCase_ ): snake_case = y_points[i] for i in range(2 ,UpperCamelCase_ ): for j in range(UpperCamelCase_ ,UpperCamelCase_ ): snake_case = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
213
1
def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): snake_case_ = generate_pascal_triangle(SCREAMING_SNAKE_CASE__ ) for row_idx in range(SCREAMING_SNAKE_CASE__ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=''' ''' ) else: print(triangle[row_idx][col_idx] , end='''''' ) print() def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) snake_case_ = [] for current_row_idx in range(SCREAMING_SNAKE_CASE__ ): snake_case_ = populate_current_row(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) triangle.append(SCREAMING_SNAKE_CASE__ ) return triangle def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): snake_case_ = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 snake_case_, snake_case_ = 1, 1 for current_col_idx in range(1 , SCREAMING_SNAKE_CASE__ ): calculate_current_element( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return current_row def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): snake_case_ = triangle[current_row_idx - 1][current_col_idx - 1] snake_case_ = triangle[current_row_idx - 1][current_col_idx] snake_case_ = above_to_left_elt + above_to_right_elt def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) snake_case_ = [[1]] for row_index in range(1 , SCREAMING_SNAKE_CASE__ ): snake_case_ = [0] + result[-1] + [0] snake_case_ = row_index + 1 # Calculate the number of distinct elements in a row snake_case_ = sum(divmod(SCREAMING_SNAKE_CASE__ , 2 ) ) snake_case_ = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] snake_case_ = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() snake_case_ = row_first_half + row_second_half result.append(SCREAMING_SNAKE_CASE__ ) return result def __SCREAMING_SNAKE_CASE (): from collections.abc import Callable from timeit import timeit def benchmark_a_function(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> None: snake_case_ = F'''{func.__name__}({value})''' snake_case_ = timeit(F'''__main__.{call}''' , setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'''{call:38} -- {timing:.4f} seconds''' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
8
'''simple docstring''' import argparse import struct import unittest class _lowerCAmelCase : def __init__(self , lowercase ): A_ : List[str] = data # Initialize hash values A_ : Tuple = [ 0X6A09_E667, 0XBB67_AE85, 0X3C6E_F372, 0XA54F_F53A, 0X510E_527F, 0X9B05_688C, 0X1F83_D9AB, 0X5BE0_CD19, ] # Initialize round constants A_ : List[Any] = [ 0X428A_2F98, 0X7137_4491, 0XB5C0_FBCF, 0XE9B5_DBA5, 0X3956_C25B, 0X59F1_11F1, 0X923F_82A4, 0XAB1C_5ED5, 0XD807_AA98, 0X1283_5B01, 0X2431_85BE, 0X550C_7DC3, 0X72BE_5D74, 0X80DE_B1FE, 0X9BDC_06A7, 0XC19B_F174, 0XE49B_69C1, 0XEFBE_4786, 0X0FC1_9DC6, 0X240C_A1CC, 0X2DE9_2C6F, 0X4A74_84AA, 0X5CB0_A9DC, 0X76F9_88DA, 0X983E_5152, 0XA831_C66D, 0XB003_27C8, 0XBF59_7FC7, 0XC6E0_0BF3, 0XD5A7_9147, 0X06CA_6351, 0X1429_2967, 0X27B7_0A85, 0X2E1B_2138, 0X4D2C_6DFC, 0X5338_0D13, 0X650A_7354, 0X766A_0ABB, 0X81C2_C92E, 0X9272_2C85, 0XA2BF_E8A1, 0XA81A_664B, 0XC24B_8B70, 0XC76C_51A3, 0XD192_E819, 0XD699_0624, 0XF40E_3585, 0X106A_A070, 0X19A4_C116, 0X1E37_6C08, 0X2748_774C, 0X34B0_BCB5, 0X391C_0CB3, 0X4ED8_AA4A, 0X5B9C_CA4F, 0X682E_6FF3, 0X748F_82EE, 0X78A5_636F, 0X84C8_7814, 0X8CC7_0208, 0X90BE_FFFA, 0XA450_6CEB, 0XBEF9_A3F7, 0XC671_78F2, ] A_ : Tuple = self.preprocessing(self.data ) self.final_hash() @staticmethod def _a (lowercase ): A_ : Any = b"""\x80""" + (b"""\x00""" * (63 - (len(lowercase ) + 8) % 64)) A_ : Optional[Any] = struct.pack(""">Q""" , (len(lowercase ) * 8) ) return data + padding + big_endian_integer def _a (self ): # Convert into blocks of 64 bytes A_ : str = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers A_ : Tuple = list(struct.unpack(""">16L""" , lowercase ) ) # add 48 0-ed integers words += [0] * 48 A_, A_, A_, A_, A_, A_, A_, A_ : Dict = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array A_ : Optional[Any] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) A_ : Tuple = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) A_ : Any = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0000_0000 # Compression A_ : Union[str, Any] = self.ror(lowercase , 6 ) ^ self.ror(lowercase , 11 ) ^ self.ror(lowercase , 25 ) A_ : List[Any] = (e & f) ^ ((~e & 0XFFFF_FFFF) & g) A_ : Dict = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0000_0000 A_ : List[str] = self.ror(lowercase , 2 ) ^ self.ror(lowercase , 13 ) ^ self.ror(lowercase , 22 ) A_ : str = (a & b) ^ (a & c) ^ (b & c) A_ : int = (sa + maj) % 0X1_0000_0000 A_, A_, A_, A_, A_, A_, A_, A_ : List[Any] = ( g, f, e, ((d + tempa) % 0X1_0000_0000), c, b, a, ((tempa + tempa) % 0X1_0000_0000), ) A_ : int = [a, b, c, d, e, f, g, h] # Modify final values A_ : Dict = [ ((element + mutated_hash_values[index]) % 0X1_0000_0000) for index, element in enumerate(self.hashes ) ] A_ : Any = """""".join([hex(lowercase )[2:].zfill(8 ) for value in self.hashes] ) def _a (self , lowercase , lowercase ): return 0XFFFF_FFFF & (value << (32 - rotations)) | (value >> rotations) class _lowerCAmelCase ( unittest.TestCase ): def _a (self ): import hashlib A_ : Optional[Any] = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(lowercase ).hash , hashlib.shaaaa(lowercase ).hexdigest() ) def a ( ): '''simple docstring''' import doctest doctest.testmod() A_ : Any = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) A_ : List[str] = parser.parse_args() A_ : Tuple = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: A_ : Union[str, Any] = f.read() else: A_ : Optional[Any] = bytes(lowerCamelCase__ , """utf-8""" ) print(SHAaaa(lowerCamelCase__ ).hash ) if __name__ == "__main__": main()
206
0
"""simple docstring""" import re def lowerCamelCase_ ( _lowerCamelCase ): if len(re.findall('[ATCG]' , _lowerCamelCase ) ) != len(_lowerCamelCase ): raise ValueError('Invalid Strand' ) return dna.translate(dna.maketrans('ATCG' , 'TAGC' ) ) if __name__ == "__main__": import doctest doctest.testmod()
316
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Dict = ['image_processor', 'tokenizer'] lowerCamelCase__ : Optional[int] = 'CLIPImageProcessor' lowerCamelCase__ : List[str] = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__(self, lowerCamelCase_=None, lowerCamelCase_=None, **lowerCamelCase_ ): '''simple docstring''' lowerCamelCase__ : str = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.', lowerCamelCase_, ) lowerCamelCase__ : int = kwargs.pop('feature_extractor' ) lowerCamelCase__ : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(lowerCamelCase_, lowerCamelCase_ ) def __call__(self, lowerCamelCase_=None, lowerCamelCase_=None, lowerCamelCase_=None, **lowerCamelCase_ ): '''simple docstring''' if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: lowerCamelCase__ : Any = self.tokenizer(lowerCamelCase_, return_tensors=lowerCamelCase_, **lowerCamelCase_ ) if images is not None: lowerCamelCase__ : List[Any] = self.image_processor(lowerCamelCase_, return_tensors=lowerCamelCase_, **lowerCamelCase_ ) if text is not None and images is not None: lowerCamelCase__ : str = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowerCamelCase_ ), tensor_type=lowerCamelCase_ ) def a__ (self, *lowerCamelCase_, **lowerCamelCase_ ): '''simple docstring''' return self.tokenizer.batch_decode(*lowerCamelCase_, **lowerCamelCase_ ) def a__ (self, *lowerCamelCase_, **lowerCamelCase_ ): '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase_, **lowerCamelCase_ ) @property def a__ (self ): '''simple docstring''' lowerCamelCase__ : str = self.tokenizer.model_input_names lowerCamelCase__ : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
316
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE__ : Optional[Any] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[Any] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : str = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys SCREAMING_SNAKE_CASE__ : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
270
import os def __magic_name__ ( ) -> str: __lowerCamelCase = os.path.join(os.path.dirname(__lowerCAmelCase ) , '''num.txt''' ) with open(__lowerCAmelCase ) as file_hand: return str(sum(int(__lowerCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
270
1
"""simple docstring""" import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class _lowerCamelCase ( unittest.TestCase ): def _lowerCAmelCase ( self : List[str] , UpperCamelCase : Optional[Any] , UpperCamelCase : str , UpperCamelCase : Tuple ) -> Tuple: """simple docstring""" self.assertEqual(len(UpperCamelCase ) , len(UpperCamelCase ) ) for a, b in zip(UpperCamelCase , UpperCamelCase ): self.assertAlmostEqual(UpperCamelCase , UpperCamelCase , delta=UpperCamelCase ) def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(UpperCamelCase ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step , 3 ) self.assertEqual(len(accumulator.gradients ) , 1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [-2.0, 5.0] , tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [0.0, 0.0] , tol=1E-2 ) def _lowerCAmelCase ( self : Dict ) -> Tuple: """simple docstring""" lowerCAmelCase__ : List[str] = None ops.enable_eager_execution_internal() lowerCAmelCase__ : Union[str, Any] = tf.config.list_physical_devices("""CPU""" ) if len(UpperCamelCase ) == 1: tf.config.set_logical_device_configuration( physical_devices[0] , [tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) lowerCAmelCase__ : Dict = tf.config.list_logical_devices(device_type="""CPU""" ) lowerCAmelCase__ : Optional[Any] = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): lowerCAmelCase__ : List[Any] = GradientAccumulator() lowerCAmelCase__ : List[str] = tf.Variable([4.0, 3.0] ) lowerCAmelCase__ : Optional[Any] = create_optimizer(5E-5 , 10 , 5 ) lowerCAmelCase__ : List[Any] = tf.Variable([0.0, 0.0] , trainable=UpperCamelCase ) def accumulate_on_replica(UpperCamelCase : Optional[int] ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients , [variable] ) ) ) @tf.function def accumulate(UpperCamelCase : List[Any] , UpperCamelCase : Tuple ): with strategy.scope(): lowerCAmelCase__ : str = strategy.experimental_local_results(UpperCamelCase ) local_variables[0].assign(UpperCamelCase ) local_variables[1].assign(UpperCamelCase ) strategy.run(UpperCamelCase , args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(UpperCamelCase ) def _check_local_values(UpperCamelCase : Optional[Any] , UpperCamelCase : str ): lowerCAmelCase__ : List[Any] = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value() , UpperCamelCase , tol=1E-2 ) self.assertListAlmostEqual(values[1].value() , UpperCamelCase , tol=1E-2 ) accumulate([1.0, 2.0] , [-1.0, 1.0] ) accumulate([3.0, -1.0] , [-1.0, -1.0] ) accumulate([-2.0, 2.0] , [3.0, -2.0] ) self.assertEqual(accumulator.step , 3 ) _check_local_values([2.0, 3.0] , [1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value() , [4.0, 3.0] , tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) _check_local_values([0.0, 0.0] , [0.0, 0.0] )
363
"""simple docstring""" from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def lowercase_ ( __UpperCAmelCase ) -> None: lowerCAmelCase__ , lowerCAmelCase__ : int = analyze_text(__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = list(""" """ + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase__ : List[str] = sum(single_char_strings.values() ) # one length string lowerCAmelCase__ : List[str] = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase__ : List[Any] = single_char_strings[ch] lowerCAmelCase__ : int = my_str / all_sum my_fir_sum += prob * math.loga(__UpperCAmelCase ) # entropy formula. # print entropy print(f"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase__ : Tuple = sum(two_char_strings.values() ) lowerCAmelCase__ : str = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase__ : Optional[int] = cha + cha if sequence in two_char_strings: lowerCAmelCase__ : int = two_char_strings[sequence] lowerCAmelCase__ : str = int(__UpperCAmelCase ) / all_sum my_sec_sum += prob * math.loga(__UpperCAmelCase ) # print second entropy print(f"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(f"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def lowercase_ ( __UpperCAmelCase ) -> tuple[dict, dict]: lowerCAmelCase__ : Any = Counter() # type: ignore lowerCAmelCase__ : Tuple = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(__UpperCAmelCase ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def lowercase_ ( ) -> Any: import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
212
0
"""simple docstring""" from string import ascii_uppercase _a = {str(ord(c) - 55): c for c in ascii_uppercase} def __a ( __lowerCamelCase, __lowerCamelCase ): if isinstance(__lowerCamelCase, __lowerCamelCase ): raise TypeError("int() can't convert non-string with explicit base" ) if num < 0: raise ValueError("parameter must be positive int" ) if isinstance(__lowerCamelCase, __lowerCamelCase ): raise TypeError("'str' object cannot be interpreted as an integer" ) if isinstance(__lowerCamelCase, __lowerCamelCase ): raise TypeError("'float' object cannot be interpreted as an integer" ) if base in (0, 1): raise ValueError("base must be >= 2" ) if base > 36: raise ValueError("base must be <= 36" ) UpperCAmelCase_ : Optional[Any] = "" UpperCAmelCase_ : int = 0 UpperCAmelCase_ : str = 0 while div != 1: UpperCAmelCase_ , UpperCAmelCase_ : str = divmod(__lowerCamelCase, __lowerCamelCase ) if base >= 11 and 9 < mod < 36: UpperCAmelCase_ : Tuple = ALPHABET_VALUES[str(__lowerCamelCase )] else: UpperCAmelCase_ : int = str(__lowerCamelCase ) new_value += actual_value UpperCAmelCase_ : Any = num // base UpperCAmelCase_ : Optional[int] = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(__lowerCamelCase ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1_000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
61
def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase )-> str: '''simple docstring''' if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) UpperCAmelCase : Dict =str(bin(__lowerCAmelCase ) ) binary_number += "0" * shift_amount return binary_number def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase )-> str: '''simple docstring''' if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) UpperCAmelCase : Any =str(bin(__lowerCAmelCase ) )[2:] if shift_amount >= len(__lowerCAmelCase ): return "0b0" UpperCAmelCase : Optional[Any] =binary_number[: len(__lowerCAmelCase ) - shift_amount] return "0b" + shifted_binary_number def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase )-> str: '''simple docstring''' if number >= 0: # Get binary representation of positive number UpperCAmelCase : Optional[Any] ='''0''' + str(bin(__lowerCAmelCase ) ).strip('''-''' )[2:] else: # Get binary (2's complement) representation of negative number UpperCAmelCase : int =len(bin(__lowerCAmelCase )[3:] ) # Find 2's complement of number UpperCAmelCase : Any =bin(abs(__lowerCAmelCase ) - (1 << binary_number_length) )[3:] UpperCAmelCase : Optional[Any] =( '''1''' + '''0''' * (binary_number_length - len(__lowerCAmelCase )) + binary_number ) if shift_amount >= len(__lowerCAmelCase ): return "0b" + binary_number[0] * len(__lowerCAmelCase ) return ( "0b" + binary_number[0] * shift_amount + binary_number[: len(__lowerCAmelCase ) - shift_amount] ) if __name__ == "__main__": import doctest doctest.testmod()
348
0
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __UpperCamelCase (_SCREAMING_SNAKE_CASE ) -> List[str]: lowercase__ = {} lowercase__ = tokenizer(example['content'] , truncation=a__ )['input_ids'] lowercase__ = len(example['content'] ) / len(output['input_ids'] ) return output lowercase_ = HfArgumentParser(PretokenizationArguments) lowercase_ = parser.parse_args() if args.num_workers is None: lowercase_ = multiprocessing.cpu_count() lowercase_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) lowercase_ = time.time() lowercase_ = load_dataset(args.dataset_name, split="""train""") print(f'''Dataset loaded in {time.time()-t_start:.2f}s''') lowercase_ = time.time() lowercase_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(f'''Dataset tokenized in {time.time()-t_start:.2f}s''') lowercase_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(f'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
371
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowercase_ = { """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
269
0
"""simple docstring""" import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase=13 ,__UpperCamelCase=7 ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=99 ,__UpperCamelCase=32 ,__UpperCamelCase=5 ,__UpperCamelCase=4 ,__UpperCamelCase=4 ,__UpperCamelCase="gelu" ,__UpperCamelCase=0.0 ,__UpperCamelCase=0.1 ,__UpperCamelCase=True ,__UpperCamelCase=512 ,__UpperCamelCase=16 ,__UpperCamelCase=2 ,__UpperCamelCase=0.02 ,__UpperCamelCase=3 ,__UpperCamelCase=4 ,__UpperCamelCase=None ,) -> int: '''simple docstring''' lowercase_ : Optional[int] = parent lowercase_ : Any = batch_size lowercase_ : Union[str, Any] = seq_length lowercase_ : Dict = is_training lowercase_ : Optional[int] = use_input_mask lowercase_ : str = use_token_type_ids lowercase_ : str = use_labels lowercase_ : List[str] = vocab_size lowercase_ : str = hidden_size lowercase_ : str = num_hidden_layers lowercase_ : str = num_attention_heads lowercase_ : Optional[Any] = intermediate_multiple_size lowercase_ : Tuple = hidden_act lowercase_ : Tuple = hidden_dropout lowercase_ : Dict = attention_dropout lowercase_ : Any = weight_tying lowercase_ : Optional[int] = max_position_embeddings lowercase_ : Tuple = type_vocab_size lowercase_ : Union[str, Any] = type_sequence_label_size lowercase_ : Union[str, Any] = initializer_range lowercase_ : Dict = num_labels lowercase_ : Optional[int] = num_choices lowercase_ : List[Any] = scope def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : Dict = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowercase_ : Optional[int] = None if self.use_input_mask: lowercase_ : int = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : Optional[int] = None if self.use_labels: lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) lowercase_ : Dict = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_multiple_size=self.intermediate_multiple_size ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,weight_tying=self.weight_tying ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=__UpperCamelCase ,initializer_range=self.initializer_range ,) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ , lowercase_ : Optional[int] = self.prepare_config_and_inputs() lowercase_ : Any = True return config, input_ids, input_mask, token_labels def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> int: '''simple docstring''' lowercase_ : Dict = GPTNeoXJapaneseModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Dict = model(__UpperCamelCase ,attention_mask=__UpperCamelCase ) lowercase_ : str = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : int = True lowercase_ : Any = GPTNeoXJapaneseModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Tuple = model(__UpperCamelCase ,attention_mask=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : Any = GPTNeoXJapaneseForCausalLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ : Optional[int] = model(__UpperCamelCase ,attention_mask=__UpperCamelCase ,labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[Any]: '''simple docstring''' lowercase_ : Any = True lowercase_ : Dict = GPTNeoXJapaneseForCausalLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() # first forward pass lowercase_ : str = model(__UpperCamelCase ,attention_mask=__UpperCamelCase ,use_cache=__UpperCamelCase ) lowercase_ : List[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase_ : Tuple = ids_tensor((self.batch_size, 3) ,config.vocab_size ) lowercase_ : Union[str, Any] = ids_tensor((self.batch_size, 3) ,vocab_size=2 ) # append to next input_ids and lowercase_ : Optional[int] = torch.cat([input_ids, next_tokens] ,dim=-1 ) lowercase_ : List[str] = torch.cat([input_mask, next_mask] ,dim=-1 ) lowercase_ : List[Any] = model(__UpperCamelCase ,attention_mask=__UpperCamelCase ,output_hidden_states=__UpperCamelCase ) lowercase_ : str = output_from_no_past['hidden_states'][0] lowercase_ : Optional[Any] = model( __UpperCamelCase ,attention_mask=__UpperCamelCase ,past_key_values=__UpperCamelCase ,output_hidden_states=__UpperCamelCase ,)['hidden_states'][0] # select random slice lowercase_ : Tuple = ids_tensor((1,) ,output_from_past.shape[-1] ).item() lowercase_ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() lowercase_ : Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCamelCase ,__UpperCamelCase ,atol=1e-3 ) ) def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ , lowercase_ : Tuple = config_and_inputs lowercase_ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( lowercase_ , lowercase_ , unittest.TestCase ): lowercase = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () lowercase = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () lowercase = ( {'feature-extraction': GPTNeoXJapaneseModel, 'text-generation': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Union[str, Any] = GPTNeoXJapaneseModelTester(self ) lowercase_ : List[Any] = ConfigTester(self ,config_class=__UpperCamelCase ,hidden_size=37 ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ , lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ , lowercase_ : int = self.model_tester.prepare_config_and_inputs_for_decoder() lowercase_ : List[str] = None self.model_tester.create_and_check_model_as_decoder(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' lowercase_ , lowercase_ , lowercase_ , lowercase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*__UpperCamelCase ) @slow def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : Dict = 'abeja/gpt-neox-japanese-2.7b' lowercase_ : str = ['データサイエンティストとは、', '100年後に必要とされる会社は、', 'フルリモートの環境で働くために必要なことは、', '国境の長いトンネルを抜けると', '美味しい日本食といえば、'] lowercase_ : int = [ 'データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。', '100年後に必要とされる会社は、「人」が中心の会社です。', 'フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。', '国境の長いトンネルを抜けると、そこは雪国だった。', '美味しい日本食といえば、やっぱりお寿司ですよね。', ] lowercase_ : Dict = GPTNeoXJapaneseTokenizer.from_pretrained(__UpperCamelCase ) lowercase_ : Dict = GPTNeoXJapaneseForCausalLM.from_pretrained(__UpperCamelCase ) lowercase_ : str = [] for prompt in prompts: lowercase_ : Tuple = tokenizer(__UpperCamelCase ,return_tensors='pt' ).input_ids lowercase_ : Optional[int] = model.generate(__UpperCamelCase ,max_length=50 ) lowercase_ : Union[str, Any] = tokenizer.batch_decode(__UpperCamelCase ,skip_special_tokens=__UpperCamelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCamelCase ,__UpperCamelCase )
213
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE =[] def lowercase__( __SCREAMING_SNAKE_CASE : list[list[int]] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int ): for i in range(len(__SCREAMING_SNAKE_CASE ) ): if board[row][i] == 1: return False for i in range(len(__SCREAMING_SNAKE_CASE ) ): if board[i][column] == 1: return False for i, j in zip(range(__SCREAMING_SNAKE_CASE , -1 , -1 ) , range(__SCREAMING_SNAKE_CASE , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(__SCREAMING_SNAKE_CASE , -1 , -1 ) , range(__SCREAMING_SNAKE_CASE , len(__SCREAMING_SNAKE_CASE ) ) ): if board[i][j] == 1: return False return True def lowercase__( __SCREAMING_SNAKE_CASE : list[list[int]] , __SCREAMING_SNAKE_CASE : int ): if row >= len(__SCREAMING_SNAKE_CASE ): solution.append(__SCREAMING_SNAKE_CASE ) printboard(__SCREAMING_SNAKE_CASE ) print() return True for i in range(len(__SCREAMING_SNAKE_CASE ) ): if is_safe(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : int = 1 solve(__SCREAMING_SNAKE_CASE , row + 1 ) lowercase_ : Dict = 0 return False def lowercase__( __SCREAMING_SNAKE_CASE : list[list[int]] ): for i in range(len(__SCREAMING_SNAKE_CASE ) ): for j in range(len(__SCREAMING_SNAKE_CASE ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) __SCREAMING_SNAKE_CASE =8 __SCREAMING_SNAKE_CASE =[[0 for i in range(n)] for j in range(n)] solve(board, 0) print("The total no. of solutions are :", len(solution))
213
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class _UpperCAmelCase( metaclass=lowerCamelCase ): lowercase__ = ['keras_nlp'] def __init__( self , *__a , **__a) -> Tuple: '''simple docstring''' requires_backends(self , ['''keras_nlp'''])
100
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ( 'This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.' 'It takes two arguments named `image` which should be the original image, and `label` which should be a text ' 'describing the elements what should be identified in the segmentation mask. The tool returns the mask.' ) lowercase__ = 'CIDAS/clipseg-rd64-refined' lowercase__ = 'image_segmenter' lowercase__ = CLIPSegForImageSegmentation lowercase__ = ['image', 'text'] lowercase__ = ['image'] def __init__( self , *__a , **__a) -> Any: '''simple docstring''' requires_backends(self , ['''vision''']) super().__init__(*__a , **__a) def UpperCAmelCase ( self , __a , __a) -> Dict: '''simple docstring''' return self.pre_processor(text=[label] , images=[image] , padding=__a , return_tensors='''pt''') def UpperCAmelCase ( self , __a) -> Tuple: '''simple docstring''' with torch.no_grad(): _UpperCamelCase = self.model(**__a).logits return logits def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = outputs.cpu().detach().numpy() _UpperCamelCase = 0 _UpperCamelCase = 1 return Image.fromarray((array * 2_55).astype(np.uinta))
100
1
"""simple docstring""" import re def A ( snake_case :str ) -> str: if len(re.findall('[ATCG]' , snake_case ) ) != len(snake_case ): raise ValueError('Invalid Strand' ) return dna.translate(dna.maketrans('ATCG' , 'TAGC' ) ) if __name__ == "__main__": import doctest doctest.testmod()
316
"""simple docstring""" import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def A ( snake_case :Union[str, Any] , snake_case :Any , snake_case :Union[str, Any] , snake_case :Any ) -> str: __UpperCamelCase = s.rsplit(snake_case , snake_case ) return new.join(snake_case ) def A ( snake_case :List[Any] ) -> int: # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if 'encoder.embeddings' not in key else 0 for key, param in state_dict.items() ) def A ( snake_case :str ) -> Union[str, Any]: __UpperCamelCase = {} __UpperCamelCase = ['group_1', 'group_2', 'group_3', 'group_4'] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: __UpperCamelCase = key.replace(f'{group_key}.' , f'{group_key}.group.' ) if "res_path" in key: __UpperCamelCase = key.replace('res_path.' , 'res_path.path.' ) if key.endswith('.w' ): __UpperCamelCase = rreplace(snake_case , '.w' , '.weight' , 1 ) if key.endswith('.b' ): __UpperCamelCase = rreplace(snake_case , '.b' , '.bias' , 1 ) __UpperCamelCase = value.float() return upgrade @torch.no_grad() def A ( snake_case :List[str] , snake_case :Tuple , snake_case :List[Any]=None , snake_case :str=True ) -> int: from dall_e import Encoder __UpperCamelCase = Encoder() if os.path.exists(snake_case ): __UpperCamelCase = torch.load(snake_case ) else: __UpperCamelCase = torch.hub.load_state_dict_from_url(snake_case ) if isinstance(snake_case , snake_case ): __UpperCamelCase = ckpt.state_dict() encoder.load_state_dict(snake_case ) if config_path is not None: __UpperCamelCase = FlavaImageCodebookConfig.from_pretrained(snake_case ) else: __UpperCamelCase = FlavaImageCodebookConfig() __UpperCamelCase = FlavaImageCodebook(snake_case ).eval() __UpperCamelCase = encoder.state_dict() __UpperCamelCase = upgrade_state_dict(snake_case ) hf_model.load_state_dict(snake_case ) __UpperCamelCase = hf_model.state_dict() __UpperCamelCase = count_parameters(snake_case ) __UpperCamelCase = count_parameters(snake_case ) assert torch.allclose(snake_case , snake_case , atol=1e-3 ) if save_checkpoint: hf_model.save_pretrained(snake_case ) else: return hf_state_dict if __name__ == "__main__": UpperCamelCase : Any = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to flava checkpoint") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") UpperCamelCase : int = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
316
1
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) def snake_case__ ( lowerCamelCase__ : Optional[Any] ) -> Any: A_ : List[str] = OrderedDict() for key, value in state_dict.items(): if key.startswith('''module.encoder''' ): A_ : Dict = key.replace('''module.encoder''' , '''glpn.encoder''' ) if key.startswith('''module.decoder''' ): A_ : Dict = key.replace('''module.decoder''' , '''decoder.stages''' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 A_ : Dict = key[key.find('''patch_embed''' ) + len('''patch_embed''' )] A_ : Tuple = key.replace(f'patch_embed{idx}' , f'patch_embeddings.{int(lowerCamelCase__ )-1}' ) if "norm" in key: A_ : Dict = key.replace('''norm''' , '''layer_norm''' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 A_ : Union[str, Any] = key[key.find('''glpn.encoder.layer_norm''' ) + len('''glpn.encoder.layer_norm''' )] A_ : Any = key.replace(f'layer_norm{idx}' , f'layer_norm.{int(lowerCamelCase__ )-1}' ) if "layer_norm1" in key: A_ : Union[str, Any] = key.replace('''layer_norm1''' , '''layer_norm_1''' ) if "layer_norm2" in key: A_ : int = key.replace('''layer_norm2''' , '''layer_norm_2''' ) if "block" in key: # replace for example block1 by block.0 A_ : Any = key[key.find('''block''' ) + len('''block''' )] A_ : Optional[int] = key.replace(f'block{idx}' , f'block.{int(lowerCamelCase__ )-1}' ) if "attn.q" in key: A_ : Any = key.replace('''attn.q''' , '''attention.self.query''' ) if "attn.proj" in key: A_ : Optional[Any] = key.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in key: A_ : List[str] = key.replace('''attn''' , '''attention.self''' ) if "fc1" in key: A_ : Tuple = key.replace('''fc1''' , '''dense1''' ) if "fc2" in key: A_ : str = key.replace('''fc2''' , '''dense2''' ) if "linear_pred" in key: A_ : List[Any] = key.replace('''linear_pred''' , '''classifier''' ) if "linear_fuse" in key: A_ : List[str] = key.replace('''linear_fuse.conv''' , '''linear_fuse''' ) A_ : str = key.replace('''linear_fuse.bn''' , '''batch_norm''' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 A_ : Union[str, Any] = key[key.find('''linear_c''' ) + len('''linear_c''' )] A_ : int = key.replace(f'linear_c{idx}' , f'linear_c.{int(lowerCamelCase__ )-1}' ) if "bot_conv" in key: A_ : str = key.replace('''bot_conv''' , '''0.convolution''' ) if "skip_conv1" in key: A_ : Dict = key.replace('''skip_conv1''' , '''1.convolution''' ) if "skip_conv2" in key: A_ : Dict = key.replace('''skip_conv2''' , '''2.convolution''' ) if "fusion1" in key: A_ : Tuple = key.replace('''fusion1''' , '''1.fusion''' ) if "fusion2" in key: A_ : Dict = key.replace('''fusion2''' , '''2.fusion''' ) if "fusion3" in key: A_ : int = key.replace('''fusion3''' , '''3.fusion''' ) if "fusion" in key and "conv" in key: A_ : List[str] = key.replace('''conv''' , '''convolutional_layer''' ) if key.startswith('''module.last_layer_depth''' ): A_ : List[str] = key.replace('''module.last_layer_depth''' , '''head.head''' ) A_ : Tuple = value return new_state_dict def snake_case__ ( lowerCamelCase__ : Optional[int] , lowerCamelCase__ : int ) -> Tuple: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) A_ : List[Any] = state_dict.pop(f'glpn.encoder.block.{i}.{j}.attention.self.kv.weight' ) A_ : List[str] = state_dict.pop(f'glpn.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict A_ : str = kv_weight[ : config.hidden_sizes[i], : ] A_ : Any = kv_bias[: config.hidden_sizes[i]] A_ : Any = kv_weight[ config.hidden_sizes[i] :, : ] A_ : List[str] = kv_bias[config.hidden_sizes[i] :] def snake_case__ ( ) -> Tuple: A_ : List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' A_ : List[Any] = Image.open(requests.get(lowerCamelCase__ , stream=lowerCamelCase__ ).raw ) return image @torch.no_grad() def snake_case__ ( lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Any , lowerCamelCase__ : int=False , lowerCamelCase__ : Union[str, Any]=None ) -> str: A_ : List[str] = GLPNConfig(hidden_sizes=[6_4, 1_2_8, 3_2_0, 5_1_2] , decoder_hidden_size=6_4 , depths=[3, 8, 2_7, 3] ) # load image processor (only resize + rescale) A_ : List[Any] = GLPNImageProcessor() # prepare image A_ : Optional[int] = prepare_img() A_ : List[Any] = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).pixel_values logger.info('''Converting model...''' ) # load original state dict A_ : List[str] = torch.load(lowerCamelCase__ , map_location=torch.device('''cpu''' ) ) # rename keys A_ : Optional[int] = rename_keys(lowerCamelCase__ ) # key and value matrices need special treatment read_in_k_v(lowerCamelCase__ , lowerCamelCase__ ) # create HuggingFace model and load state dict A_ : Tuple = GLPNForDepthEstimation(lowerCamelCase__ ) model.load_state_dict(lowerCamelCase__ ) model.eval() # forward pass A_ : Union[str, Any] = model(lowerCamelCase__ ) A_ : Optional[int] = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: A_ : Tuple = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: A_ : Union[str, Any] = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(f'Unknown model name: {model_name}' ) A_ : List[Any] = torch.Size([1, 4_8_0, 6_4_0] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ) print('''Looks ok!''' ) # finally, push to hub if required if push_to_hub: logger.info('''Pushing model and image processor to the hub...''' ) model.push_to_hub( repo_path_or_name=Path(lowerCamelCase__ , lowerCamelCase__ ) , organization='''nielsr''' , commit_message='''Add model''' , use_temp_dir=lowerCamelCase__ , ) image_processor.push_to_hub( repo_path_or_name=Path(lowerCamelCase__ , lowerCamelCase__ ) , organization='''nielsr''' , commit_message='''Add image processor''' , use_temp_dir=lowerCamelCase__ , ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) snake_case__ = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
364
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING snake_case__ = logging.get_logger(__name__) snake_case__ = { """microsoft/table-transformer-detection""": ( """https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json""" ), } class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = 'table-transformer' _lowerCAmelCase = ['past_key_values'] _lowerCAmelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Any , _lowerCamelCase : Union[str, Any]=True , _lowerCamelCase : Dict=None , _lowerCamelCase : int=3 , _lowerCamelCase : Any=100 , _lowerCamelCase : List[Any]=6 , _lowerCamelCase : Tuple=2048 , _lowerCamelCase : Any=8 , _lowerCamelCase : Dict=6 , _lowerCamelCase : Tuple=2048 , _lowerCamelCase : int=8 , _lowerCamelCase : Optional[int]=0.0 , _lowerCamelCase : List[Any]=0.0 , _lowerCamelCase : List[Any]=True , _lowerCamelCase : Optional[int]="relu" , _lowerCamelCase : Union[str, Any]=256 , _lowerCamelCase : Any=0.1 , _lowerCamelCase : Tuple=0.0 , _lowerCamelCase : Optional[int]=0.0 , _lowerCamelCase : str=0.02 , _lowerCamelCase : Tuple=1.0 , _lowerCamelCase : Dict=False , _lowerCamelCase : str="sine" , _lowerCamelCase : str="resnet50" , _lowerCamelCase : Any=True , _lowerCamelCase : List[str]=False , _lowerCamelCase : Any=1 , _lowerCamelCase : int=5 , _lowerCamelCase : Tuple=2 , _lowerCamelCase : Optional[int]=1 , _lowerCamelCase : Any=1 , _lowerCamelCase : Dict=5 , _lowerCamelCase : str=2 , _lowerCamelCase : Union[str, Any]=0.1 , **_lowerCamelCase : int , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) A_ : int = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): A_ : str = backbone_config.get('''model_type''' ) A_ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A_ : List[str] = config_class.from_dict(_lowerCamelCase ) # set timm attributes to None A_ ,A_ ,A_ : Union[str, Any] = None, None, None A_ : Optional[Any] = use_timm_backbone A_ : Optional[int] = backbone_config A_ : Optional[Any] = num_channels A_ : Dict = num_queries A_ : str = d_model A_ : List[str] = encoder_ffn_dim A_ : int = encoder_layers A_ : Optional[Any] = encoder_attention_heads A_ : List[str] = decoder_ffn_dim A_ : Any = decoder_layers A_ : List[str] = decoder_attention_heads A_ : Tuple = dropout A_ : Optional[Any] = attention_dropout A_ : Any = activation_dropout A_ : List[Any] = activation_function A_ : Dict = init_std A_ : Any = init_xavier_std A_ : List[Any] = encoder_layerdrop A_ : int = decoder_layerdrop A_ : Any = encoder_layers A_ : List[str] = auxiliary_loss A_ : List[Any] = position_embedding_type A_ : Optional[Any] = backbone A_ : Tuple = use_pretrained_backbone A_ : List[Any] = dilation # Hungarian matcher A_ : List[str] = class_cost A_ : str = bbox_cost A_ : Union[str, Any] = giou_cost # Loss coefficients A_ : Any = mask_loss_coefficient A_ : Optional[int] = dice_loss_coefficient A_ : Dict = bbox_loss_coefficient A_ : int = giou_loss_coefficient A_ : int = eos_coefficient super().__init__(is_encoder_decoder=_lowerCamelCase , **_lowerCamelCase ) @property def _a ( self : List[Any] ): """simple docstring""" return self.encoder_attention_heads @property def _a ( self : Any ): """simple docstring""" return self.d_model class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = version.parse('1.11' ) @property def _a ( self : Tuple ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def _a ( self : Optional[int] ): """simple docstring""" return 1E-5 @property def _a ( self : str ): """simple docstring""" return 12
4
0
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): a_ = 'pt' elif is_tf_available(): a_ = 'tf' else: a_ = 'jax' class UpperCAmelCase_ ( snake_case , unittest.TestCase ): UpperCamelCase =ByTaTokenizer UpperCamelCase =False def _lowerCamelCase ( self ) -> Any: super().setUp() __lowercase : List[Any] = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _lowerCamelCase ( self ) -> List[str]: return ByTaTokenizer.from_pretrained('''google/byt5-small''' ) def _lowerCamelCase ( self , **UpperCamelCase_ ) -> List[Any]: return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_=False , UpperCamelCase_=20 , UpperCamelCase_=5 ) -> Tuple: __lowercase : Optional[int] = [] for i in range(len(UpperCamelCase_ ) ): try: __lowercase : Optional[Any] = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase_ ) except UnicodeDecodeError: pass toks.append((i, tok) ) __lowercase : List[Any] = list(filter(lambda UpperCamelCase_ : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , UpperCamelCase_ ) ) __lowercase : Dict = list(filter(lambda UpperCamelCase_ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase_ ) , UpperCamelCase_ ) ) if max_length is not None and len(UpperCamelCase_ ) > max_length: __lowercase : int = toks[:max_length] if min_length is not None and len(UpperCamelCase_ ) < min_length and len(UpperCamelCase_ ) > 0: while len(UpperCamelCase_ ) < min_length: __lowercase : str = toks + toks # toks_str = [t[1] for t in toks] __lowercase : Union[str, Any] = [t[0] for t in toks] # Ensure consistency __lowercase : Tuple = tokenizer.decode(UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ ) if " " not in output_txt and len(UpperCamelCase_ ) > 1: __lowercase : int = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase_ ) + ' ' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase_ ) ) if with_prefix_space: __lowercase : List[str] = ' ' + output_txt __lowercase : Optional[Any] = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) return output_txt, output_ids def _lowerCamelCase ( self ) -> Optional[Any]: __lowercase : List[Any] = self.ta_base_tokenizer __lowercase : List[str] = tokenizer(['''hi</s>''', '''I went to the gym</s>''', '''</s>'''] ) __lowercase : str = tokenizer(['''hi''', '''I went to the gym''', ''''''] ) self.assertListEqual(batch_with_eos_added['''input_ids'''] , batch_without_eos_added['''input_ids'''] ) def _lowerCamelCase ( self ) -> Tuple: __lowercase : List[Any] = self.ta_base_tokenizer __lowercase : Tuple = 'Unicode €.' __lowercase : int = tokenizer(UpperCamelCase_ ) __lowercase : Tuple = [88, 1_13, 1_08, 1_02, 1_14, 1_03, 1_04, 35, 2_29, 1_33, 1_75, 49, 1] self.assertEqual(encoded['''input_ids'''] , UpperCamelCase_ ) # decoding __lowercase : int = tokenizer.decode(UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , '''Unicode €.</s>''' ) __lowercase : int = tokenizer('''e è é ê ë''' ) __lowercase : Dict = [1_04, 35, 1_98, 1_71, 35, 1_98, 1_72, 35, 1_98, 1_73, 35, 1_98, 1_74, 1] self.assertEqual(encoded['''input_ids'''] , UpperCamelCase_ ) # decoding __lowercase : int = tokenizer.decode(UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , '''e è é ê ë</s>''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''e è é ê ë</s>''' ) def _lowerCamelCase ( self ) -> List[str]: __lowercase : int = self.ta_base_tokenizer __lowercase : Optional[Any] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off __lowercase : str = [68, 35, 1_11, 1_14, 1_13, 1_06, 35, 1_15, 1_00, 1_17, 1_00, 1_06, 1_17, 1_00, 1_15, 1_07, 35, 1_05, 1_14, 1_17, 35, 1_18, 1_20, 1_12, 1_12, 1_00, 1_17, 1_08, 1_25, 1_00, 1_19, 1_08, 1_14, 1_13, 49, 1, 0] # fmt: on __lowercase : Dict = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors=UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) if FRAMEWORK != "jax": __lowercase : Dict = list(batch.input_ids.numpy()[0] ) else: __lowercase : int = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def _lowerCamelCase ( self ) -> str: __lowercase : Union[str, Any] = self.ta_base_tokenizer __lowercase : List[Any] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] __lowercase : Any = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors=UpperCamelCase_ ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , UpperCamelCase_ ) self.assertIn('''attention_mask''' , UpperCamelCase_ ) self.assertNotIn('''decoder_input_ids''' , UpperCamelCase_ ) self.assertNotIn('''decoder_attention_mask''' , UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Tuple: __lowercase : List[str] = self.ta_base_tokenizer __lowercase : Tuple = [ 'Summary of the text.', 'Another summary.', ] __lowercase : Any = tokenizer( text_target=UpperCamelCase_ , max_length=32 , padding='''max_length''' , truncation=UpperCamelCase_ , return_tensors=UpperCamelCase_ ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def _lowerCamelCase ( self ) -> Optional[Any]: __lowercase : Optional[Any] = self.ta_base_tokenizer __lowercase : str = ['A long paragraph for summarization. </s>'] __lowercase : Optional[Any] = ['Summary of the text. </s>'] # fmt: off __lowercase : Optional[int] = [68, 35, 1_11, 1_14, 1_13, 1_06, 35, 1_15, 1_00, 1_17, 1_00, 1_06, 1_17, 1_00, 1_15, 1_07, 35, 1_05, 1_14, 1_17, 35, 1_18, 1_20, 1_12, 1_12, 1_00, 1_17, 1_08, 1_25, 1_00, 1_19, 1_08, 1_14, 1_13, 49, 35, 1] __lowercase : Union[str, Any] = [86, 1_20, 1_12, 1_12, 1_00, 1_17, 1_24, 35, 1_14, 1_05, 35, 1_19, 1_07, 1_04, 35, 1_19, 1_04, 1_23, 1_19, 49, 35, 1] # fmt: on __lowercase : str = tokenizer(UpperCamelCase_ , text_target=UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , batch['''input_ids'''][0] ) self.assertEqual(UpperCamelCase_ , batch['''labels'''][0] ) def _lowerCamelCase ( self ) -> Dict: __lowercase : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test __lowercase : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc __lowercase : Any = tempfile.mkdtemp() __lowercase : str = ' He is very happy, UNwant\u00E9d,running' __lowercase : Any = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) tokenizer.save_pretrained(UpperCamelCase_ ) __lowercase : Dict = tokenizer.__class__.from_pretrained(UpperCamelCase_ ) __lowercase : List[Any] = after_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) shutil.rmtree(UpperCamelCase_ ) __lowercase : Optional[int] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc __lowercase : int = tempfile.mkdtemp() __lowercase : Union[str, Any] = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['''bim''', '''bambam'''] ) __lowercase : Optional[Any] = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) __lowercase : Dict = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) tokenizer.save_pretrained(UpperCamelCase_ ) __lowercase : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase_ ) __lowercase : Dict = after_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) __lowercase : str = tokenizer.__class__.from_pretrained(UpperCamelCase_ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Optional[Any]: __lowercase : Tuple = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase_ ) with open(os.path.join(UpperCamelCase_ , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: __lowercase : int = json.load(UpperCamelCase_ ) with open(os.path.join(UpperCamelCase_ , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: __lowercase : Optional[Any] = json.load(UpperCamelCase_ ) __lowercase : List[Any] = [F"""<extra_id_{i}>""" for i in range(1_25 )] __lowercase : Dict = added_tokens_extra_ids + [ 'an_additional_special_token' ] __lowercase : Optional[int] = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(UpperCamelCase_ , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(UpperCamelCase_ , UpperCamelCase_ ) with open(os.path.join(UpperCamelCase_ , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(UpperCamelCase_ , UpperCamelCase_ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files __lowercase : int = tokenizer_class.from_pretrained( UpperCamelCase_ , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained __lowercase : str = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=UpperCamelCase_ )] __lowercase : Tuple = tokenizer_class.from_pretrained( UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def _lowerCamelCase ( self ) -> List[Any]: __lowercase : int = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase_ ) __lowercase : List[str] = tokenizer_class.from_pretrained(UpperCamelCase_ ) self.assertTrue(tokenizer.decode([2_55] ) == '''''' ) def _lowerCamelCase ( self ) -> Tuple: pass def _lowerCamelCase ( self ) -> int: pass def _lowerCamelCase ( self ) -> str: pass def _lowerCamelCase ( self ) -> Optional[int]: pass def _lowerCamelCase ( self ) -> List[Any]: __lowercase : int = self.get_tokenizers(fast=UpperCamelCase_ , do_lower_case=UpperCamelCase_ ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): __lowercase : List[str] = ['t', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 'x', 't', '</s>'] __lowercase : Optional[Any] = tokenizer.convert_tokens_to_string(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Any: __lowercase : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): __lowercase : List[Any] = [ 'bos_token', 'eos_token', 'unk_token', 'sep_token', 'pad_token', 'cls_token', 'mask_token', ] __lowercase : List[Any] = 0 __lowercase : Dict = tokenizer.convert_ids_to_tokens( UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) for attr in attributes_list: setattr(UpperCamelCase_ , attr + '''_id''' , UpperCamelCase_ ) self.assertEqual(getattr(UpperCamelCase_ , UpperCamelCase_ ) , UpperCamelCase_ ) self.assertEqual(getattr(UpperCamelCase_ , attr + '''_id''' ) , UpperCamelCase_ ) setattr(UpperCamelCase_ , attr + '''_id''' , UpperCamelCase_ ) self.assertEqual(getattr(UpperCamelCase_ , UpperCamelCase_ ) , UpperCamelCase_ ) self.assertEqual(getattr(UpperCamelCase_ , attr + '''_id''' ) , UpperCamelCase_ ) setattr(UpperCamelCase_ , '''additional_special_tokens_ids''' , [] ) self.assertListEqual(getattr(UpperCamelCase_ , '''additional_special_tokens''' ) , [] ) self.assertListEqual(getattr(UpperCamelCase_ , '''additional_special_tokens_ids''' ) , [] ) setattr(UpperCamelCase_ , '''additional_special_tokens_ids''' , [token_id_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase_ , '''additional_special_tokens''' ) , [token_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase_ , '''additional_special_tokens_ids''' ) , [token_id_to_test_setters] )
249
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(__magic_name__ ) class A__ ( __magic_name__ ): def __init__( self : int , *a : Dict , **a : Union[str, Any] ): '''simple docstring''' super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _lowerCamelCase ( self : Dict , a : List[str]=None ): '''simple docstring''' lowerCAmelCase__ : Any = {} if top_k is not None: lowerCAmelCase__ : Tuple = top_k return {}, {}, postprocess_params def __call__( self : Any , a : Union[str, List[str], "Image.Image", List["Image.Image"]] , **a : List[Any] ): '''simple docstring''' return super().__call__(a , **a ) def _lowerCamelCase ( self : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = load_image(a ) lowerCAmelCase__ : Optional[int] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _lowerCamelCase ( self : Optional[int] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model(**a ) return model_outputs def _lowerCamelCase ( self : Optional[Any] , a : List[Any] , a : List[Any]=5 ): '''simple docstring''' if top_k > self.model.config.num_labels: lowerCAmelCase__ : Optional[int] = self.model.config.num_labels if self.framework == "pt": lowerCAmelCase__ : List[Any] = model_outputs.logits.softmax(-1 )[0] lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = probs.topk(a ) elif self.framework == "tf": lowerCAmelCase__ : Any = stable_softmax(model_outputs.logits , axis=-1 )[0] lowerCAmelCase__ : Any = tf.math.top_k(a , k=a ) lowerCAmelCase__ , lowerCAmelCase__ : int = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) lowerCAmelCase__ : List[Any] = scores.tolist() lowerCAmelCase__ : List[str] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
212
0
def __lowercase ( _UpperCamelCase ) ->int: """simple docstring""" if not isinstance(_UpperCamelCase, _UpperCamelCase ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1, input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
173
import os import pytest from attr import dataclass __a = '''us-east-1''' # defaults region @dataclass class __SCREAMING_SNAKE_CASE : A : str A : str = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' A : Union[str, Any] = { 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 500, 'save_steps': 5500, } A : str = {**hyperparameters, 'max_steps': 1000} @property def __lowerCamelCase ( self ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __lowerCamelCase ( self ): return f"""{self.framework}-transfromers-test""" @property def __lowerCamelCase ( self ): return f"""./tests/sagemaker/scripts/{self.framework}""" @property def __lowerCamelCase ( self ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def __lowercase ( _UpperCamelCase ) ->str: """simple docstring""" lowercase : Union[str, Any] = SageMakerTestEnvironment(framework=request.cls.framework )
173
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowercase_ ( a__ , a__ , a__ , unittest.TestCase ): __UpperCAmelCase = StableDiffusionInpaintPipeline __UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCAmelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __UpperCAmelCase = frozenset([] ) def __a ( self ): torch.manual_seed(0 ) UpperCamelCase__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a , ) UpperCamelCase__ = PNDMScheduler(skip_prk_steps=a ) torch.manual_seed(0 ) UpperCamelCase__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCamelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) UpperCamelCase__ = CLIPTextModel(a ) UpperCamelCase__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCamelCase__ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def __a ( self , a , a=0 ): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched UpperCamelCase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(a ) ).to(a ) UpperCamelCase__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCamelCase__ = Image.fromarray(np.uinta(a ) ).convert("RGB" ).resize((64, 64) ) UpperCamelCase__ = Image.fromarray(np.uinta(image + 4 ) ).convert("RGB" ).resize((64, 64) ) if str(a ).startswith("mps" ): UpperCamelCase__ = torch.manual_seed(a ) else: UpperCamelCase__ = torch.Generator(device=a ).manual_seed(a ) UpperCamelCase__ = { "prompt": "A painting of a squirrel eating a burger", "image": init_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def __a ( self ): UpperCamelCase__ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCamelCase__ = self.get_dummy_components() UpperCamelCase__ = StableDiffusionInpaintPipeline(**a ) UpperCamelCase__ = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) UpperCamelCase__ = self.get_dummy_inputs(a ) UpperCamelCase__ = sd_pipe(**a ).images UpperCamelCase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase__ = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __a ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class lowercase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench.npy" ) UpperCamelCase__ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase__ = StableDiffusionInpaintPipeline.from_pretrained(a , safety_checker=a ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing() UpperCamelCase__ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase__ = torch.manual_seed(0 ) UpperCamelCase__ = pipe( prompt=a , image=a , mask_image=a , generator=a , output_type="np" , ) UpperCamelCase__ = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 9e-3 def __a ( self ): UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench_fp16.npy" ) UpperCamelCase__ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase__ = StableDiffusionInpaintPipeline.from_pretrained( a , torch_dtype=torch.floataa , safety_checker=a , ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing() UpperCamelCase__ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase__ = torch.manual_seed(0 ) UpperCamelCase__ = pipe( prompt=a , image=a , mask_image=a , generator=a , output_type="np" , ) UpperCamelCase__ = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 5e-1 def __a ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase__ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase__ = PNDMScheduler.from_pretrained(a , subfolder="scheduler" ) UpperCamelCase__ = StableDiffusionInpaintPipeline.from_pretrained( a , safety_checker=a , scheduler=a , torch_dtype=torch.floataa , ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() UpperCamelCase__ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase__ = torch.manual_seed(0 ) UpperCamelCase__ = pipe( prompt=a , image=a , mask_image=a , generator=a , num_inference_steps=2 , output_type="np" , ) UpperCamelCase__ = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
80
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
0
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCamelCase ( lowerCAmelCase_ ): _UpperCamelCase : Optional[Any] = ['''image_processor''', '''tokenizer'''] _UpperCamelCase : Union[str, Any] = '''LayoutLMv3ImageProcessor''' _UpperCamelCase : Optional[Any] = ('''LayoutLMv3Tokenizer''', '''LayoutLMv3TokenizerFast''') def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any]=None , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Optional[Any] ) -> List[Any]: """simple docstring""" UpperCamelCase_ = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , _SCREAMING_SNAKE_CASE , ) UpperCamelCase_ = kwargs.pop("feature_extractor" ) UpperCamelCase_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __call__( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , _SCREAMING_SNAKE_CASE: Union[List[List[int]], List[List[List[int]]]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[List[int], List[List[int]]]] = None , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Union[bool, str, PaddingStrategy] = False , _SCREAMING_SNAKE_CASE: Union[bool, str, TruncationStrategy] = None , _SCREAMING_SNAKE_CASE: Optional[int] = None , _SCREAMING_SNAKE_CASE: int = 0 , _SCREAMING_SNAKE_CASE: Optional[int] = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Optional[Union[str, TensorType]] = None , **_SCREAMING_SNAKE_CASE: int , ) -> BatchEncoding: """simple docstring""" if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) # first, apply the image processor UpperCamelCase_ = self.image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase_ = [text] # add batch dimension (as the image processor always adds a batch dimension) UpperCamelCase_ = features["words"] UpperCamelCase_ = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , stride=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_token_type_ids=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE , return_overflowing_tokens=_SCREAMING_SNAKE_CASE , return_special_tokens_mask=_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , return_length=_SCREAMING_SNAKE_CASE , verbose=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) # add pixel values UpperCamelCase_ = features.pop("pixel_values" ) if return_overflowing_tokens is True: UpperCamelCase_ = self.get_overflowing_images(_SCREAMING_SNAKE_CASE , encoded_inputs["overflow_to_sample_mapping"] ) UpperCamelCase_ = images return encoded_inputs def lowercase ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict ) -> Optional[int]: """simple docstring""" UpperCamelCase_ = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" f''' {len(_SCREAMING_SNAKE_CASE )} and {len(_SCREAMING_SNAKE_CASE )}''' ) return images_with_overflow def lowercase ( self: Optional[Any] , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: Dict ) -> Union[str, Any]: """simple docstring""" return self.tokenizer.batch_decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def lowercase ( self: int , *_SCREAMING_SNAKE_CASE: str , **_SCREAMING_SNAKE_CASE: List[str] ) -> Tuple: """simple docstring""" return self.tokenizer.decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @property def lowercase ( self: Tuple ) -> str: """simple docstring""" return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def lowercase ( self: Tuple ) -> List[str]: """simple docstring""" warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , _SCREAMING_SNAKE_CASE , ) return self.image_processor_class @property def lowercase ( self: str ) -> Any: """simple docstring""" warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , _SCREAMING_SNAKE_CASE , ) return self.image_processor
328
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _UpperCAmelCase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _UpperCAmelCase = logging.getLogger() def lowerCAmelCase_ ( ) -> Optional[int]: UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument("-f" ) UpperCamelCase_ = parser.parse_args() return args.f def lowerCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_="eval" ) -> Any: UpperCamelCase_ = os.path.join(UpperCamelCase_ , F'''{split}_results.json''' ) if os.path.exists(UpperCamelCase_ ): with open(UpperCamelCase_ , "r" ) as f: return json.load(UpperCamelCase_ ) raise ValueError(F'''can\'t find {path}''' ) _UpperCAmelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _UpperCamelCase ( lowerCAmelCase_ ): def lowercase ( self: Optional[Any] ) -> Tuple: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_flax_glue.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) @slow def lowercase ( self: int ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_clm_flax.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertLess(result["eval_perplexity"] , 100 ) @slow def lowercase ( self: Any ) -> Tuple: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_summarization_flax.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE , split="test" ) self.assertGreaterEqual(result["test_rouge1"] , 10 ) self.assertGreaterEqual(result["test_rouge2"] , 2 ) self.assertGreaterEqual(result["test_rougeL"] , 7 ) self.assertGreaterEqual(result["test_rougeLsum"] , 7 ) @slow def lowercase ( self: str ) -> int: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_mlm_flax.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertLess(result["eval_perplexity"] , 42 ) @slow def lowercase ( self: Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_ta_mlm_flax.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertGreaterEqual(result["eval_accuracy"] , 0.42 ) @slow def lowercase ( self: str ) -> int: """simple docstring""" UpperCamelCase_ = 7 if get_gpu_count() > 1 else 2 UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_flax_ner.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) self.assertGreaterEqual(result["eval_f1"] , 0.3 ) @slow def lowercase ( self: Union[str, Any] ) -> Any: """simple docstring""" UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = f''' run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 '''.split() with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE ): run_qa.main() UpperCamelCase_ = get_results(_SCREAMING_SNAKE_CASE ) self.assertGreaterEqual(result["eval_f1"] , 30 ) self.assertGreaterEqual(result["eval_exact"] , 30 )
328
1
"""simple docstring""" import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( """kwargs, expected""" , [ ({"""num_shards""": 0, """max_num_jobs""": 1}, []), ({"""num_shards""": 10, """max_num_jobs""": 1}, [range(10 )]), ({"""num_shards""": 10, """max_num_jobs""": 10}, [range(UpperCamelCase_ , i + 1 ) for i in range(10 )]), ({"""num_shards""": 1, """max_num_jobs""": 10}, [range(1 )]), ({"""num_shards""": 10, """max_num_jobs""": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({"""num_shards""": 3, """max_num_jobs""": 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = _distribute_shards(**UpperCamelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, max_num_jobs, expected""" , [ ({"""foo""": 0}, 10, [{"""foo""": 0}]), ({"""shards""": [0, 1, 2, 3]}, 1, [{"""shards""": [0, 1, 2, 3]}]), ({"""shards""": [0, 1, 2, 3]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}, {"""shards""": [2]}, {"""shards""": [3]}]), ({"""shards""": [0, 1]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}]), ({"""shards""": [0, 1, 2, 3]}, 2, [{"""shards""": [0, 1]}, {"""shards""": [2, 3]}]), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = _split_gen_kwargs(UpperCamelCase_ , UpperCamelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, expected""" , [ ({"""foo""": 0}, 1), ({"""shards""": [0]}, 1), ({"""shards""": [0, 1, 2, 3]}, 4), ({"""shards""": [0, 1, 2, 3], """foo""": 0}, 4), ({"""shards""": [0, 1, 2, 3], """other""": (0, 1)}, 4), ({"""shards""": [0, 1, 2, 3], """shards2""": [0, 1]}, RuntimeError), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): if expected is RuntimeError: with pytest.raises(UpperCamelCase_ ): _number_of_shards_in_gen_kwargs(UpperCamelCase_ ) else: __SCREAMING_SNAKE_CASE = _number_of_shards_in_gen_kwargs(UpperCamelCase_ ) assert out == expected
100
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() __magic_name__ = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __magic_name__ = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.encoder.norm.weight", "encoder.layernorm.weight"), ("transformer.encoder.norm.bias", "encoder.layernorm.bias"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = state_dict.pop(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = val def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: __SCREAMING_SNAKE_CASE = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) __SCREAMING_SNAKE_CASE = value else: __SCREAMING_SNAKE_CASE = value return new_state_dict def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = """""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) __SCREAMING_SNAKE_CASE = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) __SCREAMING_SNAKE_CASE = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __SCREAMING_SNAKE_CASE = in_proj_weight[:256, :] __SCREAMING_SNAKE_CASE = in_proj_bias[:256] __SCREAMING_SNAKE_CASE = in_proj_weight[256:512, :] __SCREAMING_SNAKE_CASE = in_proj_bias[256:512] __SCREAMING_SNAKE_CASE = in_proj_weight[-256:, :] __SCREAMING_SNAKE_CASE = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention __SCREAMING_SNAKE_CASE = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) __SCREAMING_SNAKE_CASE = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __SCREAMING_SNAKE_CASE = in_proj_weight[:256, :] __SCREAMING_SNAKE_CASE = in_proj_bias[:256] __SCREAMING_SNAKE_CASE = in_proj_weight[256:512, :] __SCREAMING_SNAKE_CASE = in_proj_bias[256:512] __SCREAMING_SNAKE_CASE = in_proj_weight[-256:, :] __SCREAMING_SNAKE_CASE = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention __SCREAMING_SNAKE_CASE = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight" ) __SCREAMING_SNAKE_CASE = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias" ) # next, add query, keys and values (in that order) of cross-attention to the state dict __SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[:256, :] __SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[:256] __SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[256:512, :] __SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[256:512] __SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[-256:, :] __SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[-256:] def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = image.size __SCREAMING_SNAKE_CASE = max(UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = 800 if """detection""" in checkpoint_url else 1000 __SCREAMING_SNAKE_CASE = target_max_size / current_max_size __SCREAMING_SNAKE_CASE = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = F.to_tensor(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = F.normalize(UpperCamelCase_ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): logger.info("""Converting model...""" ) # load original state dict __SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(UpperCamelCase_ , map_location="""cpu""" ) # rename keys for src, dest in rename_keys: rename_key(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = rename_backbone_keys(UpperCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them __SCREAMING_SNAKE_CASE = """model.""" for key in state_dict.copy().keys(): if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): __SCREAMING_SNAKE_CASE = state_dict.pop(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = val # create HuggingFace model and load state dict __SCREAMING_SNAKE_CASE = TableTransformerConfig( backbone="""resnet18""" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: __SCREAMING_SNAKE_CASE = 15 __SCREAMING_SNAKE_CASE = 2 __SCREAMING_SNAKE_CASE = {0: """table""", 1: """table rotated"""} __SCREAMING_SNAKE_CASE = idalabel __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} else: __SCREAMING_SNAKE_CASE = 125 __SCREAMING_SNAKE_CASE = 6 __SCREAMING_SNAKE_CASE = { 0: """table""", 1: """table column""", 2: """table row""", 3: """table column header""", 4: """table projected row header""", 5: """table spanning cell""", } __SCREAMING_SNAKE_CASE = idalabel __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = DetrImageProcessor( format="""coco_detection""" , max_size=800 if """detection""" in checkpoint_url else 1000 ) __SCREAMING_SNAKE_CASE = TableTransformerForObjectDetection(UpperCamelCase_ ) model.load_state_dict(UpperCamelCase_ ) model.eval() # verify our conversion __SCREAMING_SNAKE_CASE = """example_pdf.png""" if """detection""" in checkpoint_url else """example_table.png""" __SCREAMING_SNAKE_CASE = hf_hub_download(repo_id="""nielsr/example-pdf""" , repo_type="""dataset""" , filename=UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = Image.open(UpperCamelCase_ ).convert("""RGB""" ) __SCREAMING_SNAKE_CASE = normalize(resize(UpperCamelCase_ , UpperCamelCase_ ) ).unsqueeze(0 ) __SCREAMING_SNAKE_CASE = model(UpperCamelCase_ ) if "detection" in checkpoint_url: __SCREAMING_SNAKE_CASE = (1, 15, 3) __SCREAMING_SNAKE_CASE = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: __SCREAMING_SNAKE_CASE = (1, 125, 7) __SCREAMING_SNAKE_CASE = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , UpperCamelCase_ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , UpperCamelCase_ , atol=1e-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(UpperCamelCase_ ).mkdir(exist_ok=UpperCamelCase_ ) model.save_pretrained(UpperCamelCase_ ) image_processor.save_pretrained(UpperCamelCase_ ) if push_to_hub: # Push model to HF hub logger.info("""Pushing model to the hub...""" ) __SCREAMING_SNAKE_CASE = ( """microsoft/table-transformer-detection""" if """detection""" in checkpoint_url else """microsoft/table-transformer-structure-recognition""" ) model.push_to_hub(UpperCamelCase_ ) image_processor.push_to_hub(UpperCamelCase_ ) if __name__ == "__main__": __magic_name__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_url", default="https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", type=str, choices=[ "https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", "https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth", ], help="URL of the Table Transformer checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __magic_name__ = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
100
1
from sklearn.metrics import mean_squared_error import datasets __snake_case = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ __snake_case = """\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. """ __snake_case = """ Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {\"raw_values\", \"uniform_average\"} or array-like of shape (n_outputs,), default=\"uniform_average\" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. \"raw_values\" : Returns a full set of errors in case of multioutput input. \"uniform_average\" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric(\"mse\") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {'mse': 0.6123724356957945} If you're using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric(\"mse\", \"multilist\") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput='raw_values') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mse': array([0.41666667, 1. ])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase__ ( datasets.Metric ): def A_ ( self : Optional[Any] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ 'https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html' ] , ) def A_ ( self : Dict ): if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value('float' ) ), "references": datasets.Sequence(datasets.Value('float' ) ), } else: return { "predictions": datasets.Value('float' ), "references": datasets.Value('float' ), } def A_ ( self : Tuple , UpperCAmelCase_ : Any , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[Any]=None , UpperCAmelCase_ : Tuple="uniform_average" , UpperCAmelCase_ : str=True ): SCREAMING_SNAKE_CASE__ = mean_squared_error( UpperCAmelCase_ , UpperCAmelCase_ , sample_weight=UpperCAmelCase_ , multioutput=UpperCAmelCase_ , squared=UpperCAmelCase_ ) return {"mse": mse}
169
import doctest from collections import deque import numpy as np class lowercase__ : def __init__( self : Optional[int] ): SCREAMING_SNAKE_CASE__ = [2, 1, 2, -1] SCREAMING_SNAKE_CASE__ = [1, 2, 3, 4] def A_ ( self : List[str] ): SCREAMING_SNAKE_CASE__ = len(self.first_signal ) SCREAMING_SNAKE_CASE__ = len(self.second_signal ) SCREAMING_SNAKE_CASE__ = max(UpperCAmelCase_ , UpperCAmelCase_ ) # create a zero matrix of max_length x max_length SCREAMING_SNAKE_CASE__ = [[0] * max_length for i in range(UpperCAmelCase_ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = deque(self.second_signal ) rotated_signal.rotate(UpperCAmelCase_ ) for j, item in enumerate(UpperCAmelCase_ ): matrix[i][j] += item # multiply the matrix with the first signal SCREAMING_SNAKE_CASE__ = np.matmul(np.transpose(UpperCAmelCase_ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(UpperCAmelCase_ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
169
1
"""simple docstring""" import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=0 ) -> List[Any]: # Format the message. if name is None: lowercase__: Any = None else: lowercase__: Optional[Any] = '''.''' * max(0 , spaces - 2 ) + '''# {:''' + str(5_0 - spaces ) + '''s}''' lowercase__: List[str] = fmt.format(__UpperCAmelCase ) # Print and recurse (if needed). if isinstance(__UpperCAmelCase , __UpperCAmelCase ): if msg is not None: print(__UpperCAmelCase ) for k in val.keys(): recursive_print(__UpperCAmelCase , val[k] , spaces + 2 ) elif isinstance(__UpperCAmelCase , torch.Tensor ): print(__UpperCAmelCase , ''':''' , val.size() ) else: print(__UpperCAmelCase , ''':''' , __UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: # Permutes layout of param tensor to [num_splits * num_heads * hidden_size, :] # for compatibility with later versions of NVIDIA Megatron-LM. # The inverse operation is performed inside Megatron-LM to read checkpoints: # https://github.com/NVIDIA/Megatron-LM/blob/v2.4/megatron/checkpointing.py#L209 # If param is the weight tensor of the self-attention block, the returned tensor # will have to be transposed one more time to be read by HuggingFace GPT2. lowercase__: str = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] lowercase__: List[str] = (num_heads, hidden_size, num_splits) + input_shape[1:] lowercase__: Optional[Any] = param.view(*__UpperCAmelCase ) lowercase__: str = param.transpose(0 , 2 ) lowercase__: Union[str, Any] = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] lowercase__: int = (num_heads, num_splits, hidden_size) + input_shape[1:] lowercase__: Optional[int] = param.view(*__UpperCAmelCase ) lowercase__: Dict = param.transpose(0 , 1 ).contiguous() lowercase__: Union[str, Any] = param.view(*__UpperCAmelCase ) return param def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> str: # The converted output model. lowercase__: int = {} # old versions did not store training args lowercase__: int = input_state_dict.get('''args''' , __UpperCAmelCase ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) lowercase__: List[str] = ds_args.padded_vocab_size lowercase__: int = ds_args.max_position_embeddings lowercase__: int = ds_args.hidden_size lowercase__: Optional[Any] = ds_args.num_layers lowercase__: int = ds_args.num_attention_heads lowercase__: Optional[Any] = ds_args.ffn_hidden_size # pprint(config) # The number of heads. lowercase__: List[Any] = config.n_head # The hidden_size per head. lowercase__: List[Any] = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): lowercase__: int = input_state_dict['''checkpoint_version'''] else: lowercase__: List[Any] = 0.0 # The model. lowercase__: Any = input_state_dict['''model'''] # The language model. lowercase__: Dict = model['''language_model'''] # The embeddings. lowercase__: str = lm['''embedding'''] # The word embeddings. lowercase__: Any = embeddings['''word_embeddings''']['''weight'''] # Truncate the embedding table to vocab_size rows. lowercase__: List[Any] = word_embeddings[: config.vocab_size, :] lowercase__: List[str] = word_embeddings # The position embeddings. lowercase__: Dict = embeddings['''position_embeddings''']['''weight'''] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] lowercase__: int = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( F"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don\'t match""" ) # Store the position embeddings. lowercase__: Optional[Any] = pos_embeddings # The transformer. lowercase__: Any = lm['''transformer'''] if '''transformer''' in lm.keys() else lm['''encoder'''] # The regex to extract layer names. lowercase__: Optional[int] = re.compile(R'''layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)''' ) # The simple map of names for "automated" rules. lowercase__: Tuple = { '''attention.dense''': '''.attn.c_proj.''', '''self_attention.dense''': '''.attn.c_proj.''', '''mlp.dense_h_to_4h''': '''.mlp.c_fc.''', '''mlp.dense_4h_to_h''': '''.mlp.c_proj.''', } # Extract the layers. for key, val in transformer.items(): # Match the name. lowercase__: List[Any] = layer_re.match(__UpperCAmelCase ) # Stop if that's not a layer if m is None: break # The index of the layer. lowercase__: List[Any] = int(m.group(1 ) ) # The name of the operation. lowercase__: Optional[int] = m.group(2 ) # Is it a weight or a bias? lowercase__: Optional[int] = m.group(3 ) # The name of the layer. lowercase__: List[Any] = F"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith('''layernorm''' ): lowercase__: Optional[int] = '''ln_1''' if op_name.startswith('''input''' ) else '''ln_2''' lowercase__: int = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. lowercase__: int = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , __UpperCAmelCase , __UpperCAmelCase ) lowercase__: List[str] = causal_mask # Insert a "dummy" tensor for masked_bias. lowercase__: List[str] = torch.tensor(-1e4 , dtype=torch.floataa ) lowercase__: int = masked_bias lowercase__: Optional[int] = fix_query_key_value_ordering(__UpperCAmelCase , __UpperCAmelCase , 3 , __UpperCAmelCase , __UpperCAmelCase ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. lowercase__: str = out_val.transpose(0 , 1 ).contiguous() # Store. lowercase__: Union[str, Any] = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": lowercase__: List[str] = fix_query_key_value_ordering(__UpperCAmelCase , __UpperCAmelCase , 3 , __UpperCAmelCase , __UpperCAmelCase ) # Store. No change of shape. lowercase__: Tuple = out_val # Transpose the weights. elif weight_or_bias == "weight": lowercase__: Any = megatron_to_transformers[op_name] lowercase__: Union[str, Any] = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": lowercase__: Optional[int] = megatron_to_transformers[op_name] lowercase__: Any = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. lowercase__: str = transformer['''final_layernorm.weight'''] lowercase__: str = transformer['''final_layernorm.bias'''] # For LM head, transformers' wants the matrix to weight embeddings. lowercase__: List[str] = word_embeddings # It should be done! return output_state_dict def SCREAMING_SNAKE_CASE__ ( ) -> Dict: # Create the argument parser. lowercase__: Dict = argparse.ArgumentParser() parser.add_argument('''--print-checkpoint-structure''' , action='''store_true''' ) parser.add_argument( '''path_to_checkpoint''' , type=__UpperCAmelCase , help='''Path to the checkpoint file (.zip archive or direct .pt file)''' , ) parser.add_argument( '''--config_file''' , default='''''' , type=__UpperCAmelCase , help='''An optional config json file describing the pre-trained model.''' , ) lowercase__: str = parser.parse_args() # Extract the basename. lowercase__: int = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(F"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith('''.zip''' ): with zipfile.ZipFile(args.path_to_checkpoint , '''r''' ) as checkpoint: with checkpoint.open('''release/mp_rank_00/model_optim_rng.pt''' ) as pytorch_dict: lowercase__: Any = torch.load(__UpperCAmelCase , map_location='''cpu''' ) else: lowercase__: List[str] = torch.load(args.path_to_checkpoint , map_location='''cpu''' ) lowercase__: Optional[int] = input_state_dict.get('''args''' , __UpperCAmelCase ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: lowercase__: Optional[Any] = '''gelu_fast''' elif ds_args.openai_gelu: lowercase__: Optional[Any] = '''gelu_new''' else: lowercase__: Dict = '''gelu''' else: # in the very early days this used to be "gelu_new" lowercase__: Optional[Any] = '''gelu_new''' # Spell out all parameters in case the defaults change. lowercase__: Dict = GPTaConfig( vocab_size=5_0_2_5_7 , n_positions=1_0_2_4 , n_embd=1_0_2_4 , n_layer=2_4 , n_head=1_6 , n_inner=4_0_9_6 , activation_function=__UpperCAmelCase , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1e-5 , initializer_range=0.0_2 , summary_type='''cls_index''' , summary_use_proj=__UpperCAmelCase , summary_activation=__UpperCAmelCase , summary_proj_to_labels=__UpperCAmelCase , summary_first_dropout=0.1 , scale_attn_weights=__UpperCAmelCase , use_cache=__UpperCAmelCase , bos_token_id=5_0_2_5_6 , eos_token_id=5_0_2_5_6 , ) else: lowercase__: Any = GPTaConfig.from_json_file(args.config_file ) lowercase__: List[str] = ['''GPT2LMHeadModel'''] # Convert. print('''Converting''' ) lowercase__: Tuple = convert_megatron_checkpoint(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(__UpperCAmelCase , __UpperCAmelCase ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: lowercase__: Dict = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": lowercase__: List[Any] = '''gpt2''' elif tokenizer_type == "PretrainedFromHF": lowercase__: Optional[Any] = ds_args.tokenizer_name_or_path else: raise ValueError(F"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: lowercase__: Dict = '''gpt2''' lowercase__: Optional[int] = AutoTokenizer.from_pretrained(__UpperCAmelCase ) lowercase__: str = type(__UpperCAmelCase ).__name__ lowercase__: Dict = tokenizer_class # Store the config to file. print('''Saving config''' ) config.save_pretrained(__UpperCAmelCase ) # Save tokenizer based on args print(F"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(__UpperCAmelCase ) # Store the state_dict to file. lowercase__: str = os.path.join(__UpperCAmelCase , '''pytorch_model.bin''' ) print(F"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(__UpperCAmelCase , __UpperCAmelCase ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
177
'''simple docstring''' def a_ ( lowerCamelCase : Optional[Any] ): return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, { 0: [6], 1: [9], 2: [4, 5], 3: [4], 4: [2, 3], 5: [2], 6: [0, 7], 7: [6], 8: [], 9: [1], }, { 0: [4], 1: [6], 2: [], 3: [5, 6, 7], 4: [0, 6], 5: [3, 8, 9], 6: [1, 3, 4, 7], 7: [3, 6, 8, 9], 8: [5, 7], 9: [5, 7], }, { 0: [1, 3], 1: [0, 2, 4], 2: [1, 3, 4], 3: [0, 2, 4], 4: [1, 2, 3], }, ][index] def a_ ( lowerCamelCase : dict[int, list[int]] ): lowerCAmelCase = 0 lowerCAmelCase = len(lowerCamelCase ) # No of vertices in graph lowerCAmelCase = [0] * n lowerCAmelCase = [False] * n def dfs(lowerCamelCase : Tuple , lowerCamelCase : str , lowerCamelCase : Dict , lowerCamelCase : str ): lowerCAmelCase = True lowerCAmelCase = id_ id_ += 1 for to in graph[at]: if to == parent: pass elif not visited[to]: dfs(lowerCamelCase , lowerCamelCase , lowerCamelCase , id_ ) lowerCAmelCase = min(low[at] , low[to] ) if id_ <= low[to]: bridges.append((at, to) if at < to else (to, at) ) else: # This edge is a back edge and cannot be a bridge lowerCAmelCase = min(low[at] , low[to] ) lowerCAmelCase = [] for i in range(lowerCamelCase ): if not visited[i]: dfs(lowerCamelCase , -1 , lowerCamelCase , id_ ) return bridges if __name__ == "__main__": import doctest doctest.testmod()
4
0
import numpy as np def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = 1e-12 , _lowerCAmelCase = 100 , ) -> List[str]: assert np.shape(A__ )[0] == np.shape(A__ )[1] # Ensure proper dimensionality. assert np.shape(A__ )[0] == np.shape(A__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(A__ ) == np.iscomplexobj(A__ ) UpperCamelCase : List[Any] = np.iscomplexobj(A__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(A__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. UpperCamelCase : Dict = False UpperCamelCase : Optional[int] = 0 UpperCamelCase : List[str] = 0 UpperCamelCase : Any = 1e12 while not convergence: # Multiple matrix by the vector. UpperCamelCase : List[Any] = np.dot(A__ , A__ ) # Normalize the resulting output vector. UpperCamelCase : List[str] = w / np.linalg.norm(A__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) UpperCamelCase : Union[str, Any] = vector.conj().T if is_complex else vector.T UpperCamelCase : List[str] = np.dot(A__ , np.dot(A__ , A__ ) ) # Check convergence. UpperCamelCase : Tuple = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: UpperCamelCase : Dict = True UpperCamelCase : Tuple = lambda_ if is_complex: UpperCamelCase : Tuple = np.real(lambda_ ) return lambda_, vector def A_ ( ) -> Dict: UpperCamelCase : Dict = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) UpperCamelCase : List[Any] = np.array([41, 4, 20] ) UpperCamelCase : Tuple = real_input_matrix.astype(np.complexaaa ) UpperCamelCase : Tuple = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T UpperCamelCase : str = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": UpperCamelCase : Optional[int] = real_input_matrix UpperCamelCase : Dict = real_vector elif problem_type == "complex": UpperCamelCase : Optional[Any] = complex_input_matrix UpperCamelCase : Any = complex_vector # Our implementation. UpperCamelCase , UpperCamelCase : Optional[Any] = power_iteration(A__ , A__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). UpperCamelCase , UpperCamelCase : Any = np.linalg.eigh(A__ ) # Last eigenvalue is the maximum one. UpperCamelCase : Dict = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. UpperCamelCase : str = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(A__ ) - np.abs(A__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
358
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A__ : def __init__( self , A_ , A_=13 , A_=30 , A_=2 , A_=3 , A_=True , A_=True , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=10 , A_=0.02 , A_=3 , A_=None , A_=2 , ): '''simple docstring''' UpperCamelCase : List[str] = parent UpperCamelCase : Tuple = batch_size UpperCamelCase : Union[str, Any] = image_size UpperCamelCase : Optional[int] = patch_size UpperCamelCase : List[str] = num_channels UpperCamelCase : Any = is_training UpperCamelCase : Dict = use_labels UpperCamelCase : List[str] = hidden_size UpperCamelCase : Dict = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : str = intermediate_size UpperCamelCase : Optional[int] = hidden_act UpperCamelCase : List[Any] = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : List[Any] = type_sequence_label_size UpperCamelCase : List[str] = initializer_range UpperCamelCase : Union[str, Any] = scope UpperCamelCase : Union[str, Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCamelCase : Optional[Any] = (image_size // patch_size) ** 2 UpperCamelCase : int = num_patches + 2 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : Tuple = None if self.use_labels: UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCamelCase( self ): '''simple docstring''' return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = TFDeiTModel(config=A_ ) UpperCamelCase : Tuple = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFDeiTForMaskedImageModeling(config=A_ ) UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCamelCase : Dict = 1 UpperCamelCase : Optional[Any] = TFDeiTForMaskedImageModeling(A_ ) UpperCamelCase : Optional[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase : Any = model(A_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.type_sequence_label_size UpperCamelCase : List[Any] = TFDeiTForImageClassification(A_ ) UpperCamelCase : Optional[int] = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase : List[Any] = 1 UpperCamelCase : Optional[Any] = TFDeiTForImageClassification(A_ ) UpperCamelCase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase : List[Any] = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase : int = config_and_inputs UpperCamelCase : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :str = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) _UpperCAmelCase :Tuple = ( { 'feature-extraction': TFDeiTModel, 'image-classification': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) _UpperCAmelCase :Dict = False _UpperCAmelCase :List[str] = False _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :Optional[int] = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFDeiTModelTester(self ) UpperCamelCase : Optional[Any] = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[int] = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCamelCase : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Dense ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : str = model_class(A_ ) UpperCamelCase : List[str] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : Optional[Any] = [*signature.parameters.keys()] UpperCamelCase : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def __UpperCamelCase( self , A_ , A_ , A_=False ): '''simple docstring''' UpperCamelCase : List[str] = super()._prepare_for_class(A_ , A_ , return_labels=A_ ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : str = TFDeiTModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def A_ ( ) -> str: UpperCamelCase : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class A__ ( unittest.TestCase ): @cached_property def __UpperCamelCase( self ): '''simple docstring''' return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFDeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ) UpperCamelCase : List[Any] = self.default_image_processor UpperCamelCase : Union[str, Any] = prepare_img() UpperCamelCase : Union[str, Any] = image_processor(images=A_ , return_tensors="tf" ) # forward pass UpperCamelCase : str = model(**A_ ) # verify the logits UpperCamelCase : Dict = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) UpperCamelCase : Tuple = tf.constant([-1.02_66, 0.19_12, -1.28_61] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , A_ , atol=1e-4 ) )
140
0
"""simple docstring""" def __magic_name__ ( lowercase , lowercase ): if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) SCREAMING_SNAKE_CASE_: Optional[int] =str(bin(lowercase ) )[2:] # remove the leading "0b" SCREAMING_SNAKE_CASE_: Any =str(bin(lowercase ) )[2:] SCREAMING_SNAKE_CASE_: Dict =max(len(lowercase ) , len(lowercase ) ) return "0b" + "".join( str(int("""1""" in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase ) , b_binary.zfill(lowercase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
173
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase = logging.get_logger(__name__) def __magic_name__ ( lowercase , lowercase=False ): SCREAMING_SNAKE_CASE_: Union[str, Any] =[] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" SCREAMING_SNAKE_CASE_: Tuple =[(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def __magic_name__ ( lowercase , lowercase , lowercase=False ): for i in range(config.num_hidden_layers ): if base_model: SCREAMING_SNAKE_CASE_: Dict ="""""" else: SCREAMING_SNAKE_CASE_: Dict ="""deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE_: Any =state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) SCREAMING_SNAKE_CASE_: Any =state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE_: int =in_proj_weight[ : config.hidden_size, : ] SCREAMING_SNAKE_CASE_: List[str] =in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE_: Optional[Any] =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE_: Optional[int] =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE_: int =in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE_: str =in_proj_bias[-config.hidden_size :] def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =dct.pop(lowercase ) SCREAMING_SNAKE_CASE_: Optional[Any] =val def __magic_name__ ( ): SCREAMING_SNAKE_CASE_: Optional[int] ="""http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE_: Dict =Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def __magic_name__ ( lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Any =DeiTConfig() # all deit models have fine-tuned heads SCREAMING_SNAKE_CASE_: Optional[int] =False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size SCREAMING_SNAKE_CASE_: Dict =1000 SCREAMING_SNAKE_CASE_: Tuple ="""huggingface/label-files""" SCREAMING_SNAKE_CASE_: int ="""imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE_: Optional[int] =json.load(open(hf_hub_download(lowercase , lowercase , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE_: Dict ={int(lowercase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_: int =idalabel SCREAMING_SNAKE_CASE_: List[str] ={v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_: Any =int(deit_name[-6:-4] ) SCREAMING_SNAKE_CASE_: int =int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): SCREAMING_SNAKE_CASE_: str =192 SCREAMING_SNAKE_CASE_: str =768 SCREAMING_SNAKE_CASE_: str =12 SCREAMING_SNAKE_CASE_: List[str] =3 elif deit_name[9:].startswith("""small""" ): SCREAMING_SNAKE_CASE_: List[str] =384 SCREAMING_SNAKE_CASE_: Optional[Any] =1536 SCREAMING_SNAKE_CASE_: Dict =12 SCREAMING_SNAKE_CASE_: List[Any] =6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): SCREAMING_SNAKE_CASE_: str =1024 SCREAMING_SNAKE_CASE_: Any =4096 SCREAMING_SNAKE_CASE_: str =24 SCREAMING_SNAKE_CASE_: str =16 # load original model from timm SCREAMING_SNAKE_CASE_: int =timm.create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model, remove and rename some keys SCREAMING_SNAKE_CASE_: List[Any] =timm_model.state_dict() SCREAMING_SNAKE_CASE_: int =create_rename_keys(lowercase , lowercase ) for src, dest in rename_keys: rename_key(lowercase , lowercase , lowercase ) read_in_q_k_v(lowercase , lowercase , lowercase ) # load HuggingFace model SCREAMING_SNAKE_CASE_: List[Any] =DeiTForImageClassificationWithTeacher(lowercase ).eval() model.load_state_dict(lowercase ) # Check outputs on an image, prepared by DeiTImageProcessor SCREAMING_SNAKE_CASE_: Tuple =int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 SCREAMING_SNAKE_CASE_: Optional[int] =DeiTImageProcessor(size=lowercase , crop_size=config.image_size ) SCREAMING_SNAKE_CASE_: List[Any] =image_processor(images=prepare_img() , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE_: Optional[Any] =encoding["""pixel_values"""] SCREAMING_SNAKE_CASE_: Tuple =model(lowercase ) SCREAMING_SNAKE_CASE_: List[str] =timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1e-3 ) Path(lowercase ).mkdir(exist_ok=lowercase ) print(f'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--deit_name""", default="""vit_deit_base_distilled_patch16_224""", type=str, help="""Name of the DeiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
173
1
'''simple docstring''' import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCAmelCase : def __init__(self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=128 , lowercase=32 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ): A_ : Union[str, Any] = parent A_ : Optional[int] = batch_size A_ : Any = seq_length A_ : int = is_training A_ : List[str] = use_input_mask A_ : Any = use_token_type_ids A_ : List[Any] = use_labels A_ : Dict = vocab_size A_ : Optional[int] = hidden_size A_ : int = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Dict = intermediate_size A_ : List[str] = hidden_act A_ : List[str] = hidden_dropout_prob A_ : Union[str, Any] = attention_probs_dropout_prob A_ : Optional[Any] = max_position_embeddings A_ : Optional[Any] = type_vocab_size A_ : List[Any] = type_sequence_label_size A_ : Tuple = initializer_range A_ : List[Any] = num_labels A_ : str = num_choices A_ : Tuple = scope def _a (self ): A_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Tuple = None if self.use_input_mask: A_ : str = random_attention_mask([self.batch_size, self.seq_length] ) A_ : Any = None if self.use_token_type_ids: A_ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Dict = None A_ : Any = None A_ : List[Any] = None if self.use_labels: A_ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : int = ids_tensor([self.batch_size] , self.num_choices ) A_ : Any = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a (self ): return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def _a (self ): ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : Union[str, Any] = self.prepare_config_and_inputs() A_ : Union[str, Any] = True A_ : List[Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Union[str, Any] = NezhaModel(config=lowercase ) model.to(lowercase ) model.eval() A_ : int = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) A_ : Optional[Any] = model(lowercase , token_type_ids=lowercase ) A_ : str = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ): A_ : Optional[int] = True A_ : Optional[Any] = NezhaModel(lowercase ) model.to(lowercase ) model.eval() A_ : Optional[int] = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , encoder_hidden_states=lowercase , encoder_attention_mask=lowercase , ) A_ : str = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , encoder_hidden_states=lowercase , ) A_ : Tuple = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Optional[Any] = NezhaForMaskedLM(config=lowercase ) model.to(lowercase ) model.eval() A_ : List[str] = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Tuple = NezhaForNextSentencePrediction(config=lowercase ) model.to(lowercase ) model.eval() A_ : Union[str, Any] = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : int = NezhaForPreTraining(config=lowercase ) model.to(lowercase ) model.eval() A_ : str = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , next_sentence_label=lowercase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Any = NezhaForQuestionAnswering(config=lowercase ) model.to(lowercase ) model.eval() A_ : Optional[int] = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , start_positions=lowercase , end_positions=lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Optional[Any] = self.num_labels A_ : int = NezhaForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() A_ : List[Any] = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : List[str] = self.num_labels A_ : Optional[int] = NezhaForTokenClassification(config=lowercase ) model.to(lowercase ) model.eval() A_ : int = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : int = self.num_choices A_ : int = NezhaForMultipleChoice(config=lowercase ) model.to(lowercase ) model.eval() A_ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ : str = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ : Optional[int] = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a (self ): A_ : Tuple = self.prepare_config_and_inputs() ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : int = config_and_inputs A_ : Any = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : Optional[Any] = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE : str = ( { 'feature-extraction': NezhaModel, 'fill-mask': NezhaForMaskedLM, 'question-answering': NezhaForQuestionAnswering, 'text-classification': NezhaForSequenceClassification, 'token-classification': NezhaForTokenClassification, 'zero-shot': NezhaForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE : List[Any] = True def _a (self , lowercase , lowercase , lowercase=False ): A_ : Optional[Any] = super()._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) if return_labels: if model_class in get_values(lowercase ): A_ : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowercase ) A_ : Tuple = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def _a (self ): A_ : Optional[int] = NezhaModelTester(self ) A_ : Any = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def _a (self ): A_ : str = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowercase ) def _a (self ): # This regression test was failing with PyTorch < 1.3 ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() A_ : str = None self.model_tester.create_and_check_model_as_decoder( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) def _a (self ): A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase ) def _a (self ): A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase ) def _a (self ): A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*lowercase ) def _a (self ): A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase ) def _a (self ): A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase ) def _a (self ): A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase ) def _a (self ): A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase ) @slow def _a (self ): for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Any = NezhaModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) @slow @require_torch_gpu def _a (self ): A_, A_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return A_ : Optional[int] = True A_ : str = model_class(config=lowercase ) A_ : str = self._prepare_for_class(lowercase , lowercase ) A_ : Tuple = torch.jit.trace( lowercase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowercase , os.path.join(lowercase , """bert.pt""" ) ) A_ : List[str] = torch.jit.load(os.path.join(lowercase , """bert.pt""" ) , map_location=lowercase ) loaded(inputs_dict["""input_ids"""].to(lowercase ) , inputs_dict["""attention_mask"""].to(lowercase ) ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): @slow def _a (self ): A_ : Dict = NezhaModel.from_pretrained("""sijunhe/nezha-cn-base""" ) A_ : List[Any] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) A_ : List[str] = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A_ : Optional[int] = model(lowercase , attention_mask=lowercase )[0] A_ : Optional[int] = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , lowercase ) A_ : List[Any] = torch.tensor([[[0.06_85, 0.24_41, 0.11_02], [0.06_00, 0.19_06, 0.13_49], [0.02_21, 0.08_19, 0.05_86]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowercase , atol=1E-4 ) ) @slow def _a (self ): A_ : str = NezhaForMaskedLM.from_pretrained("""sijunhe/nezha-cn-base""" ) A_ : Union[str, Any] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) A_ : str = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A_ : Tuple = model(lowercase , attention_mask=lowercase )[0] A_ : str = torch.Size((1, 6, 21128) ) self.assertEqual(output.shape , lowercase ) A_ : List[Any] = torch.tensor( [[-2.79_39, -1.79_02, -2.21_89], [-2.85_85, -1.89_08, -2.37_23], [-2.64_99, -1.77_50, -2.25_58]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowercase , atol=1E-4 ) )
135
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import SeqaSeqTrainer from seqaseq_training_args import SeqaSeqTrainingArguments import transformers from transformers import ( AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer, HfArgumentParser, MBartTokenizer, MBartTokenizerFast, set_seed, ) from transformers.trainer_utils import EvaluationStrategy, is_main_process from transformers.training_args import ParallelMode from utils import ( SeqaSeqDataCollator, SeqaSeqDataset, assert_all_frozen, build_compute_metrics_fn, check_output_dir, freeze_embeds, freeze_params, lmap, save_json, use_task_specific_params, write_txt_file, ) lowerCamelCase :Union[str, Any] = logging.getLogger(__name__) @dataclass class _lowerCAmelCase : __SCREAMING_SNAKE_CASE : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field( default=__UpperCAmelCase , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field( default=__UpperCAmelCase , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field( default=__UpperCAmelCase , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) __SCREAMING_SNAKE_CASE : bool = field(default=__UpperCAmelCase , metadata={'help': 'Whether tp freeze the encoder.'} ) __SCREAMING_SNAKE_CASE : bool = field(default=__UpperCAmelCase , metadata={'help': 'Whether to freeze the embeddings.'} ) @dataclass class _lowerCAmelCase : __SCREAMING_SNAKE_CASE : str = field( metadata={'help': 'The input data dir. Should contain the .tsv files (or other data files) for the task.'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field( default='summarization' , metadata={'help': 'Task name, summarization (or summarization_{dataset} for pegasus) or translation'} , ) __SCREAMING_SNAKE_CASE : Optional[int] = field( default=1_024 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __SCREAMING_SNAKE_CASE : Optional[int] = field( default=128 , metadata={ 'help': ( 'The maximum total sequence length for target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __SCREAMING_SNAKE_CASE : Optional[int] = field( default=142 , metadata={ 'help': ( 'The maximum total sequence length for validation target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded. ' 'This argument is also used to override the ``max_length`` param of ``model.generate``, which is used ' 'during ``evaluate`` and ``predict``.' ) } , ) __SCREAMING_SNAKE_CASE : Optional[int] = field( default=142 , metadata={ 'help': ( 'The maximum total sequence length for test target text after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __SCREAMING_SNAKE_CASE : Optional[int] = field(default=-1 , metadata={'help': '# training examples. -1 means use all.'} ) __SCREAMING_SNAKE_CASE : Optional[int] = field(default=-1 , metadata={'help': '# validation examples. -1 means use all.'} ) __SCREAMING_SNAKE_CASE : Optional[int] = field(default=-1 , metadata={'help': '# test examples. -1 means use all.'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field(default=__UpperCAmelCase , metadata={'help': 'Source language id for translation.'} ) __SCREAMING_SNAKE_CASE : Optional[str] = field(default=__UpperCAmelCase , metadata={'help': 'Target language id for translation.'} ) __SCREAMING_SNAKE_CASE : Optional[int] = field(default=__UpperCAmelCase , metadata={'help': '# num_beams to use for evaluation.'} ) __SCREAMING_SNAKE_CASE : bool = field( default=__UpperCAmelCase , metadata={'help': 'If only pad tokens should be ignored. This assumes that `config.pad_token_id` is defined.'} , ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' logger.info(f'***** {split} metrics *****' ) for key in sorted(metrics.keys() ): logger.info(f' {key} = {metrics[key]}' ) save_json(lowerCamelCase__ , os.path.join(lowerCamelCase__ , f'{split}_results.json' ) ) def a ( ): '''simple docstring''' A_ : Union[str, Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, SeqaSeqTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. A_, A_, A_ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: A_, A_, A_ : List[str] = parser.parse_args_into_dataclasses() check_output_dir(lowerCamelCase__ ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.parallel_mode == ParallelMode.DISTRIBUTED ) , training_args.fpaa , ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("""Training/evaluation parameters %s""" , lowerCamelCase__ ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. A_ : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) A_ : int = ("""encoder_layerdrop""", """decoder_layerdrop""", """dropout""", """attention_dropout""") for p in extra_model_params: if getattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): assert hasattr(lowerCamelCase__ , lowerCamelCase__ ), f'({config.__class__.__name__}) doesn\'t have a `{p}` attribute' setattr(lowerCamelCase__ , lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) A_ : Optional[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) A_ : Optional[Any] = AutoModelForSeqaSeqLM.from_pretrained( model_args.model_name_or_path , from_tf=""".ckpt""" in model_args.model_name_or_path , config=lowerCamelCase__ , cache_dir=model_args.cache_dir , ) # use task specific params use_task_specific_params(lowerCamelCase__ , data_args.task ) # set num_beams for evaluation if data_args.eval_beams is None: A_ : int = model.config.num_beams # set decoder_start_token_id for MBart if model.config.decoder_start_token_id is None and isinstance(lowerCamelCase__ , (MBartTokenizer, MBartTokenizerFast) ): assert ( data_args.tgt_lang is not None and data_args.src_lang is not None ), "mBart requires --tgt_lang and --src_lang" if isinstance(lowerCamelCase__ , lowerCamelCase__ ): A_ : Optional[Any] = tokenizer.lang_code_to_id[data_args.tgt_lang] else: A_ : List[str] = tokenizer.convert_tokens_to_ids(data_args.tgt_lang ) if model_args.freeze_embeds: freeze_embeds(lowerCamelCase__ ) if model_args.freeze_encoder: freeze_params(model.get_encoder() ) assert_all_frozen(model.get_encoder() ) A_ : Union[str, Any] = SeqaSeqDataset # Get datasets A_ : Union[str, Any] = ( dataset_class( lowerCamelCase__ , type_path="""train""" , data_dir=data_args.data_dir , n_obs=data_args.n_train , max_target_length=data_args.max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_train else None ) A_ : int = ( dataset_class( lowerCamelCase__ , type_path="""val""" , data_dir=data_args.data_dir , n_obs=data_args.n_val , max_target_length=data_args.val_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO else None ) A_ : Tuple = ( dataset_class( lowerCamelCase__ , type_path="""test""" , data_dir=data_args.data_dir , n_obs=data_args.n_test , max_target_length=data_args.test_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or """""" , ) if training_args.do_predict else None ) # Initialize our Trainer A_ : Optional[Any] = ( build_compute_metrics_fn(data_args.task , lowerCamelCase__ ) if training_args.predict_with_generate else None ) A_ : List[str] = SeqaSeqTrainer( model=lowerCamelCase__ , args=lowerCamelCase__ , data_args=lowerCamelCase__ , train_dataset=lowerCamelCase__ , eval_dataset=lowerCamelCase__ , data_collator=SeqaSeqDataCollator( lowerCamelCase__ , lowerCamelCase__ , model.config.decoder_start_token_id , training_args.tpu_num_cores ) , compute_metrics=lowerCamelCase__ , tokenizer=lowerCamelCase__ , ) A_ : str = {} # Training if training_args.do_train: logger.info("""*** Train ***""" ) A_ : List[Any] = trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) A_ : Any = train_result.metrics A_ : str = data_args.n_train trainer.save_model() # this also saves the tokenizer if trainer.is_world_process_zero(): handle_metrics("""train""" , lowerCamelCase__ , training_args.output_dir ) all_metrics.update(lowerCamelCase__ ) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir , """trainer_state.json""" ) ) # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) tokenizer.save_pretrained(training_args.output_dir ) # Evaluation if training_args.do_eval: logger.info("""*** Evaluate ***""" ) A_ : Tuple = trainer.evaluate(metric_key_prefix="""val""" ) A_ : str = data_args.n_val A_ : Optional[Any] = round(metrics["""val_loss"""] , 4 ) if trainer.is_world_process_zero(): handle_metrics("""val""" , lowerCamelCase__ , training_args.output_dir ) all_metrics.update(lowerCamelCase__ ) if training_args.do_predict: logger.info("""*** Predict ***""" ) A_ : Any = trainer.predict(test_dataset=lowerCamelCase__ , metric_key_prefix="""test""" ) A_ : int = test_output.metrics A_ : Tuple = data_args.n_test if trainer.is_world_process_zero(): A_ : List[Any] = round(metrics["""test_loss"""] , 4 ) handle_metrics("""test""" , lowerCamelCase__ , training_args.output_dir ) all_metrics.update(lowerCamelCase__ ) if training_args.predict_with_generate: A_ : List[Any] = tokenizer.batch_decode( test_output.predictions , skip_special_tokens=lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ ) A_ : Tuple = lmap(str.strip , lowerCamelCase__ ) write_txt_file(lowerCamelCase__ , os.path.join(training_args.output_dir , """test_generations.txt""" ) ) if trainer.is_world_process_zero(): save_json(lowerCamelCase__ , os.path.join(training_args.output_dir , """all_results.json""" ) ) return all_metrics def a ( lowerCamelCase__ ): '''simple docstring''' main() if __name__ == "__main__": main()
135
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" _snake_case = ['image_processor', 'tokenizer'] _snake_case = 'LayoutLMv3ImageProcessor' _snake_case = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ )-> Union[str, Any]: '''simple docstring''' __UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , SCREAMING_SNAKE_CASE_ , ) __UpperCamelCase = kwargs.pop('''feature_extractor''' ) __UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def __call__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = 0 , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , )-> BatchEncoding: '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.''' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''' ) # first, apply the image processor __UpperCamelCase = self.image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors=SCREAMING_SNAKE_CASE_ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): __UpperCamelCase = [text] # add batch dimension (as the image processor always adds a batch dimension) __UpperCamelCase = features['''words'''] __UpperCamelCase = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , stride=SCREAMING_SNAKE_CASE_ , pad_to_multiple_of=SCREAMING_SNAKE_CASE_ , return_token_type_ids=SCREAMING_SNAKE_CASE_ , return_attention_mask=SCREAMING_SNAKE_CASE_ , return_overflowing_tokens=SCREAMING_SNAKE_CASE_ , return_special_tokens_mask=SCREAMING_SNAKE_CASE_ , return_offsets_mapping=SCREAMING_SNAKE_CASE_ , return_length=SCREAMING_SNAKE_CASE_ , verbose=SCREAMING_SNAKE_CASE_ , return_tensors=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ , ) # add pixel values __UpperCamelCase = features.pop('''pixel_values''' ) if return_overflowing_tokens is True: __UpperCamelCase = self.get_overflowing_images(SCREAMING_SNAKE_CASE_ , encoded_inputs['''overflow_to_sample_mapping'''] ) __UpperCamelCase = images return encoded_inputs def A__ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )-> int: '''simple docstring''' __UpperCamelCase = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(SCREAMING_SNAKE_CASE_ ) != len(SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' F" {len(SCREAMING_SNAKE_CASE_ )} and {len(SCREAMING_SNAKE_CASE_ )}" ) return images_with_overflow def A__ ( self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ )-> str: '''simple docstring''' return self.tokenizer.batch_decode(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def A__ ( self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ )-> List[str]: '''simple docstring''' return self.tokenizer.decode(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) @property def A__ ( self )-> str: '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def A__ ( self )-> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , SCREAMING_SNAKE_CASE_ , ) return self.image_processor_class @property def A__ ( self )-> Any: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , SCREAMING_SNAKE_CASE_ , ) return self.image_processor
328
import math def A_ ( snake_case : int ) -> bool: '''simple docstring''' return math.sqrt(snake_case ) * math.sqrt(snake_case ) == num def A_ ( snake_case : int ) -> bool: '''simple docstring''' __UpperCamelCase = 0 __UpperCamelCase = n while left <= right: __UpperCamelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: __UpperCamelCase = mid - 1 else: __UpperCamelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
328
1
"""simple docstring""" import requests from bsa import BeautifulSoup def __a ( __lowerCamelCase = "https://www.worldometers.info/coronavirus" ): UpperCAmelCase_ : Tuple = BeautifulSoup(requests.get(__lowerCamelCase ).text, "html.parser" ) UpperCAmelCase_ : Optional[Any] = soup.findAll("h1" ) UpperCAmelCase_ : Union[str, Any] = soup.findAll("div", {"class": "maincounter-number"} ) keys += soup.findAll("span", {"class": "panel-title"} ) values += soup.findAll("div", {"class": "number-table-main"} ) return {key.text.strip(): value.text.strip() for key, value in zip(__lowerCamelCase, __lowerCamelCase )} if __name__ == "__main__": print('\033[1m' + 'COVID-19 Status of the World' + '\033[0m\n') for key, value in world_covidaa_stats().items(): print(f"""{key}\n{value}\n""")
23
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : List[str] = [file for file in os.listdir(lowercase_ ) if os.path.isfile(os.path.join(lowercase_ , lowercase_ ) )] if identifier is not None: UpperCAmelCase_ : Dict = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase_ , lowercase_ ): for n_ in n_identifier: UpperCAmelCase_ : str = [file for file in files if n_ not in file] else: UpperCAmelCase_ : Any = [file for file in files if n_identifier not in file] UpperCAmelCase_ : Union[str, Any] = ignore_files or [] ignore_files.append("__init__.py" ) UpperCAmelCase_ : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowercase_ ) if only_modules: UpperCAmelCase_ : str = file.split("." )[0] try: UpperCAmelCase_ : str = getattr(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = doctest.DocTestSuite(lowercase_ ) UpperCAmelCase_ : int = unittest.TextTestRunner().run(lowercase_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase_ : Optional[Any] = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Path("src/transformers" ) UpperCAmelCase_ : str = "modeling" UpperCAmelCase_ : Optional[Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowercase_ , identifier=lowercase_ , ignore_files=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = Path("src/transformers" ) UpperCAmelCase_ : Any = "tokenization" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = "configuration" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowercase_ , n_identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = Path("docs/source" ) UpperCAmelCase_ : Union[str, Any] = ["favicon.ico"] self.analyze_directory(lowercase_ , ignore_files=lowercase_ , only_modules=lowercase_ )
23
1
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _lowerCAmelCase : List[Any] = random.Random() def lowerCAmelCase ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1.0 , _lowerCAmelCase : Any=None , _lowerCAmelCase : Union[str, Any]=None ): """simple docstring""" if rng is None: UpperCAmelCase__ = global_rng UpperCAmelCase__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _UpperCamelCase ( unittest.TestCase ): def __init__( self :Dict , lowerCamelCase :Optional[int] , lowerCamelCase :List[str]=7 , lowerCamelCase :int=400 , lowerCamelCase :Dict=2000 , lowerCamelCase :List[Any]=1 , lowerCamelCase :Tuple=0.0 , lowerCamelCase :Tuple=1_6000 , lowerCamelCase :Tuple=True , lowerCamelCase :Dict=True , ) -> int: UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = min_seq_length UpperCAmelCase__ = max_seq_length UpperCAmelCase__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCAmelCase__ = feature_size UpperCAmelCase__ = padding_value UpperCAmelCase__ = sampling_rate UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = do_normalize def UpperCAmelCase_ ( self :int ) -> Union[str, Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCAmelCase_ ( self :List[Any] , lowerCamelCase :str=False , lowerCamelCase :Tuple=False ) -> str: def _flatten(lowerCamelCase :Union[str, Any] ): return list(itertools.chain(*lowerCamelCase ) ) if equal_length: UpperCAmelCase__ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size UpperCAmelCase__ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCAmelCase__ = [np.asarray(lowerCamelCase ) for x in speech_inputs] return speech_inputs class _UpperCamelCase ( lowerCAmelCase , unittest.TestCase ): UpperCAmelCase_ = WavaVecaFeatureExtractor def UpperCAmelCase_ ( self :List[str] ) -> List[Any]: UpperCAmelCase__ = WavaVecaFeatureExtractionTester(self ) def UpperCAmelCase_ ( self :Optional[int] , lowerCamelCase :int ) -> Tuple: self.assertTrue(np.all(np.mean(lowerCamelCase , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCAmelCase_ ( self :Tuple ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCAmelCase__ = [np.asarray(lowerCamelCase ) for speech_input in speech_inputs] # Test not batched input UpperCAmelCase__ = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values UpperCAmelCase__ = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) # Test batched UpperCAmelCase__ = feat_extract(lowerCamelCase , return_tensors="np" ).input_values UpperCAmelCase__ = feat_extract(lowerCamelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase , lowerCamelCase ): self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCAmelCase__ = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCAmelCase__ = np.asarray(lowerCamelCase ) UpperCAmelCase__ = feat_extract(lowerCamelCase , return_tensors="np" ).input_values UpperCAmelCase__ = feat_extract(lowerCamelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase , lowerCamelCase ): self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) def UpperCAmelCase_ ( self :Dict ) -> Tuple: UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCAmelCase__ = ["longest", "max_length", "do_not_pad"] UpperCAmelCase__ = [None, 1600, None] for max_length, padding in zip(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = feat_extract(lowerCamelCase , padding=lowerCamelCase , max_length=lowerCamelCase , return_tensors="np" ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCAmelCase_ ( self :List[str] ) -> Any: UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = range(800 , 1400 , 200 ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in lengths] UpperCAmelCase__ = ["longest", "max_length", "do_not_pad"] UpperCAmelCase__ = [None, 1600, None] for max_length, padding in zip(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = feat_extract(lowerCamelCase , max_length=lowerCamelCase , padding=lowerCamelCase ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCAmelCase_ ( self :List[str] ) -> Any: UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=1000 , padding="max_length" , return_tensors="np" ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCAmelCase_ ( self :Optional[Any] ) -> int: UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=1000 , padding="longest" , return_tensors="np" ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=2000 , padding="longest" , return_tensors="np" ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCAmelCase_ ( self :List[str] ) -> Optional[int]: import torch UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = np.random.rand(100 ).astype(np.floataa ) UpperCAmelCase__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCAmelCase__ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) UpperCAmelCase__ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCAmelCase_ ( self :Any ) -> Optional[int]: # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: UpperCAmelCase__ = WavaVecaConfig.from_pretrained(lowerCamelCase ) UpperCAmelCase__ = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == "layer" )
169
from ...processing_utils import ProcessorMixin class _UpperCamelCase ( lowerCAmelCase ): UpperCAmelCase_ = ["""image_processor""", """feature_extractor"""] UpperCAmelCase_ = """TvltImageProcessor""" UpperCAmelCase_ = """TvltFeatureExtractor""" def __init__( self :List[str] , lowerCamelCase :Dict , lowerCamelCase :Tuple ) -> Any: super().__init__(image_processor=lowerCamelCase , feature_extractor=lowerCamelCase ) UpperCAmelCase__ = image_processor UpperCAmelCase__ = feature_extractor def __call__( self :Union[str, Any] , lowerCamelCase :List[str]=None , lowerCamelCase :int=None , lowerCamelCase :List[Any]=None , lowerCamelCase :Dict=None , lowerCamelCase :List[str]=False , lowerCamelCase :Optional[Any]=False , *lowerCamelCase :List[Any] , **lowerCamelCase :Dict , ) -> List[str]: if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) UpperCAmelCase__ = None if images is not None: UpperCAmelCase__ = self.image_processor(lowerCamelCase , mask_pixel=lowerCamelCase , *lowerCamelCase , **lowerCamelCase ) if images_mixed is not None: UpperCAmelCase__ = self.image_processor(lowerCamelCase , is_mixed=lowerCamelCase , *lowerCamelCase , **lowerCamelCase ) if audio is not None: UpperCAmelCase__ = self.feature_extractor( lowerCamelCase , *lowerCamelCase , sampling_rate=lowerCamelCase , mask_audio=lowerCamelCase , **lowerCamelCase ) UpperCAmelCase__ = {} if audio is not None: output_dict.update(lowerCamelCase ) if images is not None: output_dict.update(lowerCamelCase ) if images_mixed_dict is not None: output_dict.update(lowerCamelCase ) return output_dict @property def UpperCAmelCase_ ( self :Dict ) -> Optional[Any]: UpperCAmelCase__ = self.image_processor.model_input_names UpperCAmelCase__ = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
169
1
from __future__ import annotations import math class __A : def __init__( self , UpperCAmelCase_ ): lowerCamelCase =size # approximate the overall size of segment tree with given value lowerCamelCase =[0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCamelCase =[0 for i in range(0 , 4 * size )] lowerCamelCase =[0 for i in range(0 , 4 * size )] # flag for lazy update def _snake_case ( self , UpperCAmelCase_ ): return idx * 2 def _snake_case ( self , UpperCAmelCase_ ): return idx * 2 + 1 def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if left_element == right_element: lowerCamelCase =a[left_element - 1] else: lowerCamelCase =(left_element + right_element) // 2 self.build(self.left(UpperCAmelCase_ ) , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) self.build(self.right(UpperCAmelCase_ ) , mid + 1 , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCamelCase =max( self.segment_tree[self.left(UpperCAmelCase_ )] , self.segment_tree[self.right(UpperCAmelCase_ )] ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if self.flag[idx] is True: lowerCamelCase =self.lazy[idx] lowerCamelCase =False if left_element != right_element: lowerCamelCase =self.lazy[idx] lowerCamelCase =self.lazy[idx] lowerCamelCase =True lowerCamelCase =True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCamelCase =val if left_element != right_element: lowerCamelCase =val lowerCamelCase =val lowerCamelCase =True lowerCamelCase =True return True lowerCamelCase =(left_element + right_element) // 2 self.update(self.left(UpperCAmelCase_ ) , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) self.update(self.right(UpperCAmelCase_ ) , mid + 1 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCamelCase =max( self.segment_tree[self.left(UpperCAmelCase_ )] , self.segment_tree[self.right(UpperCAmelCase_ )] ) return True def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if self.flag[idx] is True: lowerCamelCase =self.lazy[idx] lowerCamelCase =False if left_element != right_element: lowerCamelCase =self.lazy[idx] lowerCamelCase =self.lazy[idx] lowerCamelCase =True lowerCamelCase =True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCamelCase =(left_element + right_element) // 2 lowerCamelCase =self.query(self.left(UpperCAmelCase_ ) , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCamelCase =self.query(self.right(UpperCAmelCase_ ) , mid + 1 , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return max(UpperCAmelCase_ , UpperCAmelCase_ ) def __str__( self ): return str([self.query(1 , 1 , self.size , UpperCAmelCase_ , UpperCAmelCase_ ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": UpperCAmelCase__ : List[str] =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] UpperCAmelCase__ : Optional[int] =15 UpperCAmelCase__ : int =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 1_11) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 2_35) print(segt)
262
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging UpperCAmelCase__ : List[Any] =logging.get_logger(__name__) UpperCAmelCase__ : Dict ={'''vocab_file''': '''spiece.model'''} UpperCAmelCase__ : Dict ={ '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } UpperCAmelCase__ : List[str] ={ '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) UpperCAmelCase__ : Any =0 UpperCAmelCase__ : List[Any] =1 UpperCAmelCase__ : Union[str, Any] =2 UpperCAmelCase__ : Tuple =3 UpperCAmelCase__ : int =4 class __A ( a ): __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = """left""" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<sep>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<cls>" , UpperCAmelCase_="<mask>" , UpperCAmelCase_=["<eop>", "<eod>"] , UpperCAmelCase_ = None , **UpperCAmelCase_ , ): # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase =AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token lowerCamelCase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase_ , remove_space=UpperCAmelCase_ , keep_accents=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase_ , ) lowerCamelCase =3 lowerCamelCase =do_lower_case lowerCamelCase =remove_space lowerCamelCase =keep_accents lowerCamelCase =vocab_file lowerCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase_ ) @property def _snake_case ( self ): return len(self.sp_model ) def _snake_case ( self ): lowerCamelCase ={self.convert_ids_to_tokens(UpperCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowerCamelCase =self.__dict__.copy() lowerCamelCase =None return state def __setstate__( self , UpperCAmelCase_ ): lowerCamelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase ={} lowerCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _snake_case ( self , UpperCAmelCase_ ): if self.remove_space: lowerCamelCase =""" """.join(inputs.strip().split() ) else: lowerCamelCase =inputs lowerCamelCase =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: lowerCamelCase =unicodedata.normalize("""NFKD""" , UpperCAmelCase_ ) lowerCamelCase ="""""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase_ )] ) if self.do_lower_case: lowerCamelCase =outputs.lower() return outputs def _snake_case ( self , UpperCAmelCase_ ): lowerCamelCase =self.preprocess_text(UpperCAmelCase_ ) lowerCamelCase =self.sp_model.encode(UpperCAmelCase_ , out_type=UpperCAmelCase_ ) lowerCamelCase =[] for piece in pieces: if len(UpperCAmelCase_ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowerCamelCase =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase_ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase =cur_pieces[1:] else: lowerCamelCase =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase_ ) else: new_pieces.append(UpperCAmelCase_ ) return new_pieces def _snake_case ( self , UpperCAmelCase_ ): return self.sp_model.PieceToId(UpperCAmelCase_ ) def _snake_case ( self , UpperCAmelCase_ ): return self.sp_model.IdToPiece(UpperCAmelCase_ ) def _snake_case ( self , UpperCAmelCase_ ): lowerCamelCase ="""""".join(UpperCAmelCase_ ).replace(UpperCAmelCase_ , """ """ ).strip() return out_string def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , **UpperCAmelCase_ , ): lowerCamelCase =kwargs.pop("""use_source_tokenizer""" , UpperCAmelCase_ ) lowerCamelCase =self.convert_ids_to_tokens(UpperCAmelCase_ , skip_special_tokens=UpperCAmelCase_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCamelCase =[] lowerCamelCase =[] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase_ ) ) lowerCamelCase =[] sub_texts.append(UpperCAmelCase_ ) else: current_sub_text.append(UpperCAmelCase_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase_ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowerCamelCase ="""""".join(UpperCAmelCase_ ) lowerCamelCase =( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCamelCase =self.clean_up_tokenization(UpperCAmelCase_ ) return clean_text else: return text def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCamelCase =[self.sep_token_id] lowerCamelCase =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase_ )) + [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] return ([0] * len(UpperCAmelCase_ )) + [1, 1] def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCamelCase =[self.sep_token_id] lowerCamelCase =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase =os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase_ , """wb""" ) as fi: lowerCamelCase =self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase_ ) return (out_vocab_file,)
262
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class _A : def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : int=13 , __SCREAMING_SNAKE_CASE : Optional[Any]=7 , __SCREAMING_SNAKE_CASE : Dict=False , __SCREAMING_SNAKE_CASE : Optional[int]=True , __SCREAMING_SNAKE_CASE : Tuple=False , __SCREAMING_SNAKE_CASE : Optional[int]=False , __SCREAMING_SNAKE_CASE : str=19 , __SCREAMING_SNAKE_CASE : str=32 , __SCREAMING_SNAKE_CASE : List[str]=5 , __SCREAMING_SNAKE_CASE : Union[str, Any]=4 , __SCREAMING_SNAKE_CASE : Dict=37 , __SCREAMING_SNAKE_CASE : Optional[int]="gelu" , __SCREAMING_SNAKE_CASE : Dict=0.1 , __SCREAMING_SNAKE_CASE : str=0.1 , __SCREAMING_SNAKE_CASE : str=512 , __SCREAMING_SNAKE_CASE : Dict=16 , __SCREAMING_SNAKE_CASE : Union[str, Any]=2 , __SCREAMING_SNAKE_CASE : List[Any]=0.02 , __SCREAMING_SNAKE_CASE : int=3 , __SCREAMING_SNAKE_CASE : Optional[Any]=4 , __SCREAMING_SNAKE_CASE : Optional[Any]=None , ): '''simple docstring''' __a = parent __a = batch_size __a = seq_length __a = is_training __a = use_input_mask __a = use_token_type_ids __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_size __a = type_sequence_label_size __a = initializer_range __a = num_labels __a = num_choices __a = scope def _lowerCamelCase ( self : str): '''simple docstring''' __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length]) __a = None __a = None __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size) __a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __a = ids_tensor([self.batch_size] , self.num_choices) __a = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = EsmConfig( vocab_size=33 , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , is_folding_model=__SCREAMING_SNAKE_CASE , esmfold_config={'''trunk''': {'''num_blocks''': 2}, '''fp16_esm''': False} , ) return config def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : List[str]): '''simple docstring''' __a = EsmForProteinFolding(config=__SCREAMING_SNAKE_CASE).float() model.to(__SCREAMING_SNAKE_CASE) model.eval() __a = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE) __a = model(__SCREAMING_SNAKE_CASE) __a = model(__SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.positions.shape , (8, self.batch_size, self.seq_length, 14, 3)) self.parent.assertEqual(result.angles.shape , (8, self.batch_size, self.seq_length, 7, 2)) def _lowerCamelCase ( self : int): '''simple docstring''' __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _A ( __UpperCAmelCase ,__UpperCAmelCase ,unittest.TestCase ): UpperCamelCase__ : Tuple = False UpperCamelCase__ : Optional[Any] = (EsmForProteinFolding,) if is_torch_available() else () UpperCamelCase__ : List[str] = () UpperCamelCase__ : str = {} if is_torch_available() else {} UpperCamelCase__ : Optional[Any] = False def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = EsmFoldModelTester(self) __a = ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , hidden_size=37) def _lowerCamelCase ( self : Dict): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE) @unittest.skip('''Does not support attention outputs''') def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' pass @unittest.skip def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' pass @unittest.skip('''Esm does not support embedding resizing''') def _lowerCamelCase ( self : List[str]): '''simple docstring''' pass @unittest.skip('''Esm does not support embedding resizing''') def _lowerCamelCase ( self : List[str]): '''simple docstring''' pass @unittest.skip('''ESMFold does not support passing input embeds!''') def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' pass @unittest.skip('''ESMFold does not support head pruning.''') def _lowerCamelCase ( self : int): '''simple docstring''' pass @unittest.skip('''ESMFold does not support head pruning.''') def _lowerCamelCase ( self : Dict): '''simple docstring''' pass @unittest.skip('''ESMFold does not support head pruning.''') def _lowerCamelCase ( self : int): '''simple docstring''' pass @unittest.skip('''ESMFold does not support head pruning.''') def _lowerCamelCase ( self : Tuple): '''simple docstring''' pass @unittest.skip('''ESMFold does not support head pruning.''') def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' pass @unittest.skip('''ESMFold does not output hidden states in the normal way.''') def _lowerCamelCase ( self : Tuple): '''simple docstring''' pass @unittest.skip('''ESMfold does not output hidden states in the normal way.''') def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' pass @unittest.skip('''ESMFold only has one output format.''') def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' pass @unittest.skip('''This test doesn\'t work for ESMFold and doesn\'t test core functionality''') def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' pass @unittest.skip('''ESMFold does not support input chunking.''') def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' pass @unittest.skip('''ESMFold doesn\'t respect you and it certainly doesn\'t respect your initialization arguments.''') def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''') def _lowerCamelCase ( self : int): '''simple docstring''' pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''') def _lowerCamelCase ( self : List[str]): '''simple docstring''' pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''') def _lowerCamelCase ( self : Any): '''simple docstring''' pass @unittest.skip('''ESMFold doesn\'t support data parallel.''') def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''') def _lowerCamelCase ( self : str): '''simple docstring''' pass @require_torch class _A ( __UpperCAmelCase ): @slow def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = EsmForProteinFolding.from_pretrained('''facebook/esmfold_v1''').float() model.eval() __a = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]]) __a = model(__SCREAMING_SNAKE_CASE)['''positions'''] __a = torch.tensor([2.58_28, 0.79_93, -10.93_34] , dtype=torch.floataa) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0] , __SCREAMING_SNAKE_CASE , atol=1E-4))
49
from string import ascii_lowercase, ascii_uppercase def UpperCamelCase ( __lowercase : str ): '''simple docstring''' if not sentence: return "" A_ : List[str] = dict(zip(__lowercase ,__lowercase ) ) return lower_to_upper.get(sentence[0] ,sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
140
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings UpperCAmelCase__ : Tuple = R'\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information.\n\n Args:\n title_sep (`str`, *optional*, defaults to `" / "`):\n Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`].\n doc_sep (`str`, *optional*, defaults to `" // "`):\n Separator inserted between the text of the retrieved document and the original input when calling\n [`RagRetriever`].\n n_docs (`int`, *optional*, defaults to 5):\n Number of documents to retrieve.\n max_combined_length (`int`, *optional*, defaults to 300):\n Max length of contextualized input returned by [`~RagRetriever.__call__`].\n retrieval_vector_size (`int`, *optional*, defaults to 768):\n Dimensionality of the document embeddings indexed by [`RagRetriever`].\n retrieval_batch_size (`int`, *optional*, defaults to 8):\n Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated\n [`RagRetriever`].\n dataset (`str`, *optional*, defaults to `"wiki_dpr"`):\n A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids\n using `datasets.list_datasets()`).\n dataset_split (`str`, *optional*, defaults to `"train"`)\n Which split of the `dataset` to load.\n index_name (`str`, *optional*, defaults to `"compressed"`)\n The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and\n `"compressed"`.\n index_path (`str`, *optional*)\n The path to the serialized faiss index on disk.\n passages_path (`str`, *optional*):\n A path to text passages compatible with the faiss index. Required if using\n [`~models.rag.retrieval_rag.LegacyIndex`]\n use_dummy_dataset (`bool`, *optional*, defaults to `False`)\n Whether to load a "dummy" variant of the dataset specified by `dataset`.\n label_smoothing (`float`, *optional*, defaults to 0.0):\n Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing\n in the loss calculation. If set to 0, no label smoothing is performed.\n do_marginalize (`bool`, *optional*, defaults to `False`):\n If `True`, the logits are marginalized over all documents by making use of\n `torch.nn.functional.log_softmax`.\n reduce_loss (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation.\n do_deduplication (`bool`, *optional*, defaults to `True`):\n Whether or not to deduplicate the generations from different context documents for a given input. Has to be\n set to `False` if used while training with distributed backend.\n exclude_bos_score (`bool`, *optional*, defaults to `False`):\n Whether or not to disregard the BOS token when computing the loss.\n output_retrieved(`bool`, *optional*, defaults to `False`):\n If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and\n `context_attention_mask` are returned. See returned tensors for more detail.\n use_cache (`bool`, *optional*, defaults to `True`):\n Whether or not the model should return the last key/values attentions (not used by all models).\n forced_eos_token_id (`int`, *optional*):\n The id of the token to force as the last generated token when `max_length` is reached. Usually set to\n `eos_token_id`.\n' @add_start_docstrings(SCREAMING_SNAKE_CASE__ ) class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : int = '''rag''' __UpperCamelCase : List[Any] = True def __init__( self : List[str] , lowerCAmelCase_ : Optional[Any]=None , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : List[Any]=None , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : List[str]=" / " , lowerCAmelCase_ : int=" // " , lowerCAmelCase_ : List[str]=5 , lowerCAmelCase_ : Union[str, Any]=3_0_0 , lowerCAmelCase_ : Union[str, Any]=7_6_8 , lowerCAmelCase_ : int=8 , lowerCAmelCase_ : Optional[int]="wiki_dpr" , lowerCAmelCase_ : int="train" , lowerCAmelCase_ : Optional[Any]="compressed" , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : str=None , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Optional[int]=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : str=None , **lowerCAmelCase_ : Optional[int] , ): """simple docstring""" super().__init__( bos_token_id=lowerCAmelCase_ , pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , prefix=lowerCAmelCase_ , vocab_size=lowerCAmelCase_ , **lowerCAmelCase_ , ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _A: Optional[Any] = kwargs.pop('''question_encoder''' ) _A: Optional[int] = question_encoder_config.pop('''model_type''' ) _A: Tuple = kwargs.pop('''generator''' ) _A: int = decoder_config.pop('''model_type''' ) from ..auto.configuration_auto import AutoConfig _A: List[Any] = AutoConfig.for_model(lowerCAmelCase_ , **lowerCAmelCase_ ) _A: Tuple = AutoConfig.for_model(lowerCAmelCase_ , **lowerCAmelCase_ ) _A: List[str] = reduce_loss _A: Tuple = label_smoothing _A: Optional[Any] = exclude_bos_score _A: List[str] = do_marginalize _A: Dict = title_sep _A: List[Any] = doc_sep _A: Any = n_docs _A: Any = max_combined_length _A: Dict = dataset _A: int = dataset_split _A: int = index_name _A: Optional[Any] = retrieval_vector_size _A: Union[str, Any] = retrieval_batch_size _A: List[Any] = passages_path _A: List[str] = index_path _A: Dict = use_dummy_dataset _A: int = output_retrieved _A: Any = do_deduplication _A: str = use_cache if self.forced_eos_token_id is None: _A: List[str] = getattr(self.generator , '''forced_eos_token_id''' , lowerCAmelCase_ ) @classmethod def __magic_name__ ( cls : str , lowerCAmelCase_ : PretrainedConfig , lowerCAmelCase_ : PretrainedConfig , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" return cls(question_encoder=question_encoder_config.to_dict() , generator=generator_config.to_dict() , **lowerCAmelCase_ ) def __magic_name__ ( self : Dict ): """simple docstring""" _A: Optional[int] = copy.deepcopy(self.__dict__ ) _A: Tuple = self.question_encoder.to_dict() _A: Any = self.generator.to_dict() _A: Optional[int] = self.__class__.model_type return output
301
# This is the module that test_patching.py uses to test patch_submodule() import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests UpperCAmelCase__ : str = open # noqa: we just need to have a builtin inside this module to test it properly
301
1
"""simple docstring""" from __future__ import annotations from dataclasses import dataclass @dataclass class _snake_case : snake_case__ = 42 snake_case__ = None snake_case__ = None def lowercase_ ( _lowerCamelCase: TreeNode | None ) -> bool: '''simple docstring''' def is_valid_tree(_lowerCamelCase: TreeNode | None ) -> bool: if node is None: return True if not isinstance(_lowerCamelCase , _lowerCamelCase ): return False try: float(node.data ) except (TypeError, ValueError): return False return is_valid_tree(node.left ) and is_valid_tree(node.right ) if not is_valid_tree(_lowerCamelCase ): raise ValueError( "Each node should be type of TreeNode and data should be float." ) def is_binary_search_tree_recursive_check( _lowerCamelCase: TreeNode | None , _lowerCamelCase: float , _lowerCamelCase: float ) -> bool: if node is None: return True return ( left_bound < node.data < right_bound and is_binary_search_tree_recursive_check(node.left , _lowerCamelCase , node.data ) and is_binary_search_tree_recursive_check( node.right , node.data , _lowerCamelCase ) ) return is_binary_search_tree_recursive_check(_lowerCamelCase , -float("inf" ) , float("inf" ) ) if __name__ == "__main__": import doctest doctest.testmod()
135
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) __A = { '''configuration_layoutlmv3''': [ '''LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LayoutLMv3Config''', '''LayoutLMv3OnnxConfig''', ], '''processing_layoutlmv3''': ['''LayoutLMv3Processor'''], '''tokenization_layoutlmv3''': ['''LayoutLMv3Tokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ['''LayoutLMv3TokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ '''LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LayoutLMv3ForQuestionAnswering''', '''LayoutLMv3ForSequenceClassification''', '''LayoutLMv3ForTokenClassification''', '''LayoutLMv3Model''', '''LayoutLMv3PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ '''TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLayoutLMv3ForQuestionAnswering''', '''TFLayoutLMv3ForSequenceClassification''', '''TFLayoutLMv3ForTokenClassification''', '''TFLayoutLMv3Model''', '''TFLayoutLMv3PreTrainedModel''', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ['''LayoutLMv3FeatureExtractor'''] __A = ['''LayoutLMv3ImageProcessor'''] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys __A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
135
1
'''simple docstring''' import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class a ( _a , _a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = IFPipeline SCREAMING_SNAKE_CASE : Optional[int] = TEXT_TO_IMAGE_PARAMS - {"width", "height", "latents"} SCREAMING_SNAKE_CASE : Optional[int] = TEXT_TO_IMAGE_BATCH_PARAMS SCREAMING_SNAKE_CASE : str = PipelineTesterMixin.required_optional_params - {"latents"} def lowerCamelCase__ ( self : Tuple ) -> Optional[Any]: return self._get_dummy_components() def lowerCamelCase__ ( self : Optional[Any] , snake_case : Optional[int] , snake_case : List[Any]=0 ) -> Optional[Any]: if str(snake_case ).startswith('''mps''' ): __UpperCAmelCase : Optional[Any] = torch.manual_seed(snake_case ) else: __UpperCAmelCase : Dict = torch.Generator(device=snake_case ).manual_seed(snake_case ) __UpperCAmelCase : str = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def lowerCamelCase__ ( self : List[str] ) -> Optional[int]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def lowerCamelCase__ ( self : Dict ) -> Union[str, Any]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def lowerCamelCase__ ( self : List[str] ) -> Tuple: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def lowerCamelCase__ ( self : Any ) -> List[str]: self._test_save_load_local() def lowerCamelCase__ ( self : Union[str, Any] ) -> str: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowerCamelCase__ ( self : str ) -> Optional[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self : Dict ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self : Any ) -> Tuple: # if __UpperCAmelCase : Union[str, Any] = IFPipeline.from_pretrained('''DeepFloyd/IF-I-XL-v1.0''' , variant='''fp16''' , torch_dtype=torch.floataa ) __UpperCAmelCase : List[str] = IFSuperResolutionPipeline.from_pretrained( '''DeepFloyd/IF-II-L-v1.0''' , variant='''fp16''' , torch_dtype=torch.floataa , text_encoder=snake_case , tokenizer=snake_case ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to('''cuda''' ) __UpperCAmelCase : Tuple = pipe_a.encode_prompt('''anime turtle''' , device='''cuda''' ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() __UpperCAmelCase : Any = None __UpperCAmelCase : Optional[Any] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(snake_case , snake_case , snake_case , snake_case ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img __UpperCAmelCase : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) __UpperCAmelCase : int = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(snake_case , snake_case , snake_case , snake_case ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting __UpperCAmelCase : List[str] = IFInpaintingPipeline(**pipe_a.components ) __UpperCAmelCase : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(snake_case , snake_case , snake_case , snake_case ) def lowerCamelCase__ ( self : List[str] , snake_case : Any , snake_case : Union[str, Any] , snake_case : Optional[int] , snake_case : str ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() __UpperCAmelCase : Dict = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : List[str] = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , num_inference_steps=2 , generator=snake_case , output_type='''np''' , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (64, 64, 3) __UpperCAmelCase : List[str] = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 __UpperCAmelCase : Optional[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) # pipeline 2 _start_torch_memory_measurement() __UpperCAmelCase : int = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : List[Any] = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , image=snake_case , generator=snake_case , num_inference_steps=2 , output_type='''np''' , ) __UpperCAmelCase : Optional[Any] = output.images[0] assert image.shape == (256, 256, 3) __UpperCAmelCase : List[str] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 __UpperCAmelCase : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) def lowerCamelCase__ ( self : Optional[int] , snake_case : Optional[int] , snake_case : Union[str, Any] , snake_case : str , snake_case : Dict ) -> str: # pipeline 1 _start_torch_memory_measurement() __UpperCAmelCase : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : Dict = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : Dict = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , image=snake_case , num_inference_steps=2 , generator=snake_case , output_type='''np''' , ) __UpperCAmelCase : int = output.images[0] assert image.shape == (64, 64, 3) __UpperCAmelCase : Any = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 __UpperCAmelCase : Optional[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) # pipeline 2 _start_torch_memory_measurement() __UpperCAmelCase : Optional[int] = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : Optional[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : List[str] = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , image=snake_case , original_image=snake_case , generator=snake_case , num_inference_steps=2 , output_type='''np''' , ) __UpperCAmelCase : int = output.images[0] assert image.shape == (256, 256, 3) __UpperCAmelCase : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 __UpperCAmelCase : Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) def lowerCamelCase__ ( self : str , snake_case : Union[str, Any] , snake_case : Optional[int] , snake_case : Optional[Any] , snake_case : Optional[Any] ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() __UpperCAmelCase : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(snake_case ) __UpperCAmelCase : Optional[Any] = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : Dict = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , image=snake_case , mask_image=snake_case , num_inference_steps=2 , generator=snake_case , output_type='''np''' , ) __UpperCAmelCase : Union[str, Any] = output.images[0] assert image.shape == (64, 64, 3) __UpperCAmelCase : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 __UpperCAmelCase : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) # pipeline 2 _start_torch_memory_measurement() __UpperCAmelCase : int = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : Any = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(snake_case ) __UpperCAmelCase : Tuple = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(snake_case ) __UpperCAmelCase : Union[str, Any] = pipe_a( prompt_embeds=snake_case , negative_prompt_embeds=snake_case , image=snake_case , mask_image=snake_case , original_image=snake_case , generator=snake_case , num_inference_steps=2 , output_type='''np''' , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (256, 256, 3) __UpperCAmelCase : Any = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 __UpperCAmelCase : str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(snake_case , snake_case ) def _a ( ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
350
'''simple docstring''' import requests __UpperCAmelCase :Union[str, Any] = "https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=" def _a ( _lowercase : str ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page['''articles'''] , 1 ): print(F'{i}.) {article["title"]}' ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key="<Your BBC News API key goes here>")
240
0
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class SCREAMING_SNAKE_CASE: """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = None # Automatically constructed lowerCamelCase__ = "dict" lowerCamelCase__ = None lowerCamelCase__ = field(default="""Translation""" , init=A__ , repr=A__ ) def __call__( self : Any ) -> Optional[Any]: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def A ( self : List[Any] ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value('''string''' ) for k in sorted(self.languages )} @dataclass class SCREAMING_SNAKE_CASE: """simple docstring""" lowerCamelCase__ = None lowerCamelCase__ = None lowerCamelCase__ = None # Automatically constructed lowerCamelCase__ = "dict" lowerCamelCase__ = None lowerCamelCase__ = field(default="""TranslationVariableLanguages""" , init=A__ , repr=A__ ) def A ( self : List[Any] ) -> List[str]: UpperCAmelCase : str = sorted(set(self.languages ) ) if self.languages else None UpperCAmelCase : Dict = len(self.languages ) if self.languages else None def __call__( self : int ) -> str: return pa.struct({'''language''': pa.list_(pa.string() ), '''translation''': pa.list_(pa.string() )} ) def A ( self : Tuple , __snake_case : int ) -> Any: UpperCAmelCase : List[str] = set(self.languages ) if self.languages and set(__snake_case ) - lang_set: raise ValueError( F"""Some languages in example ({", ".join(sorted(set(__snake_case ) - lang_set ) )}) are not in valid set ({", ".join(__snake_case )}).""" ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. UpperCAmelCase : List[Any] = [] for lang, text in translation_dict.items(): if isinstance(__snake_case , __snake_case ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. UpperCAmelCase , UpperCAmelCase : Optional[Any] = zip(*sorted(__snake_case ) ) return {"language": languages, "translation": translations} def A ( self : Tuple ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value('''string''' ) ), "translation": Sequence(Value('''string''' ) ), }
23
'''simple docstring''' import random from .binary_exp_mod import bin_exp_mod def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1000 ) -> int: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd UpperCAmelCase : str = n - 1 UpperCAmelCase : List[Any] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) UpperCAmelCase : List[str] = 0 while count < prec: UpperCAmelCase : int = random.randint(2 , n - 1 ) UpperCAmelCase : List[str] = bin_exp_mod(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if b != 1: UpperCAmelCase : int = True for _ in range(_lowerCAmelCase ): if b == n - 1: UpperCAmelCase : Dict = False break UpperCAmelCase : str = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCamelCase__: Optional[int] = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
23
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging UpperCAmelCase_ : int = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = ["input_features"] def __init__( self : int , lowercase_ : str=80 , lowercase_ : Any=16000 , lowercase_ : Any=160 , lowercase_ : int=30 , lowercase_ : Any=400 , lowercase_ : int=0.0 , lowercase_ : Union[str, Any]=False , **lowercase_ : Optional[int] , ): '''simple docstring''' super().__init__( feature_size=lowercase_ , sampling_rate=lowercase_ , padding_value=lowercase_ , return_attention_mask=lowercase_ , **lowercase_ , ) SCREAMING_SNAKE_CASE_ : Tuple = n_fft SCREAMING_SNAKE_CASE_ : Tuple = hop_length SCREAMING_SNAKE_CASE_ : Optional[Any] = chunk_length SCREAMING_SNAKE_CASE_ : int = chunk_length * sampling_rate SCREAMING_SNAKE_CASE_ : Tuple = self.n_samples // hop_length SCREAMING_SNAKE_CASE_ : List[Any] = sampling_rate SCREAMING_SNAKE_CASE_ : Tuple = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=lowercase_ , min_frequency=0.0 , max_frequency=80_00.0 , sampling_rate=lowercase_ , norm='''slaney''' , mel_scale='''slaney''' , ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowercase_ : np.array): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = spectrogram( lowercase_ , window_function(self.n_fft , '''hann''') , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = log_spec[:, :-1] SCREAMING_SNAKE_CASE_ : Dict = np.maximum(lowercase_ , log_spec.max() - 8.0) SCREAMING_SNAKE_CASE_ : Tuple = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def _SCREAMING_SNAKE_CASE ( lowercase_ : List[np.ndarray] , lowercase_ : List[np.ndarray] , lowercase_ : float = 0.0): '''simple docstring''' if attention_mask is not None: SCREAMING_SNAKE_CASE_ : Union[str, Any] = np.array(lowercase_ , np.intaa) SCREAMING_SNAKE_CASE_ : List[Any] = [] for vector, length in zip(lowercase_ , attention_mask.sum(-1)): SCREAMING_SNAKE_CASE_ : int = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7) if length < normed_slice.shape[0]: SCREAMING_SNAKE_CASE_ : Any = padding_value normed_input_values.append(lowercase_) else: SCREAMING_SNAKE_CASE_ : List[Any] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7) for x in input_values] return normed_input_values def __call__( self : List[Any] , lowercase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowercase_ : bool = True , lowercase_ : Optional[int] = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[str] = "max_length" , lowercase_ : Optional[int] = None , lowercase_ : Optional[int] = None , lowercase_ : Optional[bool] = None , **lowercase_ : Optional[int] , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.') else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''') SCREAMING_SNAKE_CASE_ : Tuple = isinstance(lowercase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}') SCREAMING_SNAKE_CASE_ : str = is_batched_numpy or ( isinstance(lowercase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: SCREAMING_SNAKE_CASE_ : Optional[Any] = [np.asarray([speech] , dtype=np.floataa).T for speech in raw_speech] elif not is_batched and not isinstance(lowercase_ , np.ndarray): SCREAMING_SNAKE_CASE_ : Tuple = np.asarray(lowercase_ , dtype=np.floataa) elif isinstance(lowercase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): SCREAMING_SNAKE_CASE_ : Any = raw_speech.astype(np.floataa) # always return batch if not is_batched: SCREAMING_SNAKE_CASE_ : str = [np.asarray([raw_speech]).T] SCREAMING_SNAKE_CASE_ : int = BatchFeature({'''input_features''': raw_speech}) # convert into correct format for padding SCREAMING_SNAKE_CASE_ : Any = self.pad( lowercase_ , padding=lowercase_ , max_length=max_length if max_length else self.n_samples , truncation=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: SCREAMING_SNAKE_CASE_ : Dict = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) SCREAMING_SNAKE_CASE_ : Dict = np.stack(padded_inputs['''input_features'''] , axis=0) # make sure list is in array format SCREAMING_SNAKE_CASE_ : str = padded_inputs.get('''input_features''').transpose(2 , 0 , 1) SCREAMING_SNAKE_CASE_ : List[str] = [self._np_extract_fbank_features(lowercase_) for waveform in input_features[0]] if isinstance(input_features[0] , lowercase_): SCREAMING_SNAKE_CASE_ : str = [np.asarray(lowercase_ , dtype=np.floataa) for feature in input_features] else: SCREAMING_SNAKE_CASE_ : Any = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) SCREAMING_SNAKE_CASE_ : Tuple = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: SCREAMING_SNAKE_CASE_ : Optional[int] = padded_inputs.convert_to_tensors(lowercase_) return padded_inputs def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = copy.deepcopy(self.__dict__) SCREAMING_SNAKE_CASE_ : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
318
"""simple docstring""" import argparse import os import re import packaging.version UpperCAmelCase_ : Any = """examples/""" UpperCAmelCase_ : Optional[int] = { """examples""": (re.compile(r"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(r"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(r"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), r"""\1version=\"VERSION\","""), """doc""": (re.compile(r"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } UpperCAmelCase_ : List[Any] = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } UpperCAmelCase_ : Optional[int] = """README.md""" def _A (__a , __a , __a ) -> int: """simple docstring""" with open(__a , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: SCREAMING_SNAKE_CASE_ : Optional[Any] = f.read() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[Any] = REPLACE_PATTERNS[pattern] SCREAMING_SNAKE_CASE_ : Optional[int] = replace.replace('''VERSION''' , __a ) SCREAMING_SNAKE_CASE_ : Tuple = re_pattern.sub(__a , __a ) with open(__a , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(__a ) def _A (__a ) -> int: """simple docstring""" for folder, directories, fnames in os.walk(__a ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__a , __a ) , __a , pattern='''examples''' ) def _A (__a , __a=False ) -> List[str]: """simple docstring""" for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__a , __a , __a ) if not patch: update_version_in_examples(__a ) def _A () -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = '''🤗 Transformers currently provides the following architectures''' SCREAMING_SNAKE_CASE_ : Optional[int] = '''1. Want to contribute a new model?''' with open(__a , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: SCREAMING_SNAKE_CASE_ : Tuple = f.readlines() # Find the start of the list. SCREAMING_SNAKE_CASE_ : Tuple = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 SCREAMING_SNAKE_CASE_ : Dict = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): SCREAMING_SNAKE_CASE_ : List[Any] = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(__a , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__a ) def _A () -> List[str]: """simple docstring""" with open(REPLACE_FILES['''init'''] , '''r''' ) as f: SCREAMING_SNAKE_CASE_ : Any = f.read() SCREAMING_SNAKE_CASE_ : Dict = REPLACE_PATTERNS['''init'''][0].search(__a ).groups()[0] return packaging.version.parse(__a ) def _A (__a=False ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: SCREAMING_SNAKE_CASE_ : List[Any] = default_version.base_version elif patch: SCREAMING_SNAKE_CASE_ : int = f'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: SCREAMING_SNAKE_CASE_ : Any = f'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. SCREAMING_SNAKE_CASE_ : int = input(f'Which version are you releasing? [{default_version}]' ) if len(__a ) == 0: SCREAMING_SNAKE_CASE_ : Optional[Any] = default_version print(f'Updating version to {version}.' ) global_version_update(__a , patch=__a ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def _A () -> Any: """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = get_version() SCREAMING_SNAKE_CASE_ : Any = f'{current_version.major}.{current_version.minor + 1}.0.dev0' SCREAMING_SNAKE_CASE_ : Union[str, Any] = current_version.base_version # Check with the user we got that right. SCREAMING_SNAKE_CASE_ : int = input(f'Which version are we developing now? [{dev_version}]' ) if len(__a ) == 0: SCREAMING_SNAKE_CASE_ : Optional[int] = dev_version print(f'Updating version to {version}.' ) global_version_update(__a ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": UpperCAmelCase_ : Optional[int] = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") UpperCAmelCase_ : int = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
318
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : List[Any] =logging.get_logger(__name__) _UpperCAmelCase : List[str] ={ """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/config.json""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/config.json""", """funnel-transformer/medium-base""": """https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json""", """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/config.json""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json""", """funnel-transformer/xlarge-base""": """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = """funnel""" SCREAMING_SNAKE_CASE__ : Dict = { """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", } def __init__( self , __lowercase=3_0_5_2_2 , __lowercase=[4, 4, 4] , __lowercase=None , __lowercase=2 , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=6_4 , __lowercase=3_0_7_2 , __lowercase="gelu_new" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=None , __lowercase=1e-9 , __lowercase="mean" , __lowercase="relative_shift" , __lowercase=True , __lowercase=True , __lowercase=True , **__lowercase , ) -> Optional[Any]: lowerCAmelCase_ : int = vocab_size lowerCAmelCase_ : Union[str, Any] = block_sizes lowerCAmelCase_ : List[str] = [1] * len(__lowercase ) if block_repeats is None else block_repeats assert len(__lowercase ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." lowerCAmelCase_ : Optional[Any] = num_decoder_layers lowerCAmelCase_ : List[Any] = d_model lowerCAmelCase_ : List[Any] = n_head lowerCAmelCase_ : Dict = d_head lowerCAmelCase_ : List[Any] = d_inner lowerCAmelCase_ : List[str] = hidden_act lowerCAmelCase_ : Dict = hidden_dropout lowerCAmelCase_ : List[str] = attention_dropout lowerCAmelCase_ : int = activation_dropout lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Tuple = initializer_std lowerCAmelCase_ : Optional[int] = layer_norm_eps assert pooling_type in [ "mean", "max", ], f"""Got {pooling_type} for `pooling_type` but only 'mean' and 'max' are supported.""" lowerCAmelCase_ : Dict = pooling_type assert attention_type in [ "relative_shift", "factorized", ], f"""Got {attention_type} for `attention_type` but only 'relative_shift' and 'factorized' are supported.""" lowerCAmelCase_ : Dict = attention_type lowerCAmelCase_ : Dict = separate_cls lowerCAmelCase_ : Any = truncate_seq lowerCAmelCase_ : Any = pool_q_only super().__init__(**__lowercase ) @property def lowercase_ ( self ) -> List[str]: return sum(self.block_sizes ) @num_hidden_layers.setter def lowercase_ ( self , __lowercase ) -> str: raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.''' ) @property def lowercase_ ( self ) -> List[Any]: return len(self.block_sizes ) @num_blocks.setter def lowercase_ ( self , __lowercase ) -> Dict: raise NotImplementedError('''This model does not support the setting of `num_blocks`. Please set `block_sizes`.''' )
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
from __future__ import annotations from math import pow, sqrt def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance == 0: return {"resistance": sqrt(pow(a_ , 2 ) - pow(a_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(a_ , 2 ) - pow(a_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(a_ , 2 ) + pow(a_ , 2 ) )} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
367
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=SCREAMING_SNAKE_CASE__ ) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = field(default='''summarization''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowerCamelCase__ = Features({'''text''': Value('''string''' )} ) lowerCamelCase__ = Features({'''summary''': Value('''string''' )} ) lowerCamelCase__ = "text" lowerCamelCase__ = "summary" @property def __A ( self : Dict ) -> Dict[str, str]: return {self.text_column: "text", self.summary_column: "summary"}
305
0
"""simple docstring""" import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate SCREAMING_SNAKE_CASE_ = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow('''''', '''|''', '''|'''), datarow=DataRow('''''', '''|''', '''|'''), padding=1, with_header_hide=None, ) SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = {'''type''': '''section''', '''text''': {'''type''': '''plain_text''', '''text''': '''No failed tests! 🤗''', '''emoji''': True}} SCREAMING_SNAKE_CASE_ = [ { '''type''': '''header''', '''text''': { '''type''': '''plain_text''', '''text''': F"🤗 Accelerate nightly {os.environ.get('TEST_TYPE', '')} test results", '''emoji''': True, }, } ] SCREAMING_SNAKE_CASE_ = 0 for log in Path().glob('''*.log'''): SCREAMING_SNAKE_CASE_ = 0 with open(log, '''r''') as f: for line in f: SCREAMING_SNAKE_CASE_ = json.loads(line) if line.get('''nodeid''', '''''') != "": SCREAMING_SNAKE_CASE_ = line['''nodeid'''] if line.get('''duration''', None) is not None: SCREAMING_SNAKE_CASE_ = F"{line['duration']:.4f}" if line.get('''outcome''', '''''') == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split('''_''')[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) SCREAMING_SNAKE_CASE_ = [] log.unlink() SCREAMING_SNAKE_CASE_ = '''''' SCREAMING_SNAKE_CASE_ = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += F"*{name[1:]}: {num_failed} failed test*\n" else: message += F"*{name[1:]}: {num_failed} failed tests*\n" SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = {} for test in failed_tests: SCREAMING_SNAKE_CASE_ = test[0].split('''::''') SCREAMING_SNAKE_CASE_ = data[0].split('''/''')[-1] if data[0] not in filesafailed: SCREAMING_SNAKE_CASE_ = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) SCREAMING_SNAKE_CASE_ = [test[0] for test in failed_table] SCREAMING_SNAKE_CASE_ = list(set(files)) # Count number of instances in failed_tests SCREAMING_SNAKE_CASE_ = [] for file in individual_files: table.append([file, len(filesafailed[file])]) SCREAMING_SNAKE_CASE_ = tabulate( table, headers=['''Test Location''', '''Num Failed'''], tablefmt=hf_table_format, stralign='''right''', ) message += F"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3_000: SCREAMING_SNAKE_CASE_ = '''Too many failed tests, please see the full report in the Action results.''' SCREAMING_SNAKE_CASE_ = len(err) + 10 SCREAMING_SNAKE_CASE_ = message[: 3_000 - offset] + F"\n...\n```\n{err}" print(F"### {message}") else: SCREAMING_SNAKE_CASE_ = '''No failed tests! 🤗''' print(F"## {message}") payload.append(no_error_payload) if os.environ.get('''TEST_TYPE''', '''''') != "": from slack_sdk import WebClient SCREAMING_SNAKE_CASE_ = WebClient(token=os.environ['''SLACK_API_TOKEN''']) if message != "No failed tests! 🤗": SCREAMING_SNAKE_CASE_ = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': message, }, } payload.append(md_report) SCREAMING_SNAKE_CASE_ = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': '''*For more details:*''', }, '''accessory''': { '''type''': '''button''', '''text''': { '''type''': '''plain_text''', '''text''': '''Check Action results''', '''emoji''': True, }, '''url''': F"https://github.com/{os.environ['GITHUB_REPOSITORY']}/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } payload.append(action_button) SCREAMING_SNAKE_CASE_ = { '''type''': '''context''', '''elements''': [ { '''type''': '''plain_text''', '''text''': F"Nightly {os.environ.get('TEST_TYPE')} test results for {date.today()}", } ], } payload.append(date_report) SCREAMING_SNAKE_CASE_ = client.chat_postMessage(channel='''#accelerate-ci-daily''', text=message, blocks=payload) SCREAMING_SNAKE_CASE_ = response.data['''ts'''] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name SCREAMING_SNAKE_CASE_ = '''''' for i, row in enumerate(test_failures): if row[0] != test_class: SCREAMING_SNAKE_CASE_ = row[0] else: SCREAMING_SNAKE_CASE_ = '''''' SCREAMING_SNAKE_CASE_ = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': F"Test location: {test_location}\n```\n{tabulate(test_failures, headers=['Class', 'Test'], tablefmt=hf_table_format, stralign='right')}\n```", }, } client.chat_postMessage( channel='''#accelerate-ci-daily''', thread_ts=ts, blocks=[payload], )
301
"""simple docstring""" import os from distutils.util import strtobool def lowercase (_lowerCAmelCase , _lowerCAmelCase ): for e in env_keys: __lowerCAmelCase = int(os.environ.get(_lowerCAmelCase , -1 ) ) if val >= 0: return val return default def lowercase (_lowerCAmelCase , _lowerCAmelCase=False ): __lowerCAmelCase = os.environ.get(_lowerCAmelCase , str(_lowerCAmelCase ) ) return strtobool(_lowerCAmelCase ) == 1 # As its name indicates `strtobool` actually returns an int... def lowercase (_lowerCAmelCase , _lowerCAmelCase="no" ): __lowerCAmelCase = os.environ.get(_lowerCAmelCase , str(_lowerCAmelCase ) ) return value
301
1
from __future__ import annotations from math import pow, sqrt def _a ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float ): """simple docstring""" if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(__SCREAMING_SNAKE_CASE , 2 ) - pow(__SCREAMING_SNAKE_CASE , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(__SCREAMING_SNAKE_CASE , 2 ) - pow(__SCREAMING_SNAKE_CASE , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(__SCREAMING_SNAKE_CASE , 2 ) + pow(__SCREAMING_SNAKE_CASE , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
370
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def _a ( SCREAMING_SNAKE_CASE : Optional[Any]=None , SCREAMING_SNAKE_CASE : int=None ): """simple docstring""" return field(default_factory=lambda: default , metadata=SCREAMING_SNAKE_CASE ) @dataclass class __magic_name__ : A: str = field( metadata={"help": "The csv file to plot."} , ) A: bool = field( default=__lowerCAmelCase , metadata={"help": "Whether to plot along batch size or sequence length. Defaults to sequence length."} , ) A: bool = field( default=__lowerCAmelCase , metadata={"help": "Whether the csv file has time results or memory results. Defaults to memory results."} , ) A: bool = field( default=__lowerCAmelCase , metadata={"help": "Disable logarithmic scale when plotting"} , ) A: bool = field( default=__lowerCAmelCase , metadata={ "help": "Whether the csv file has training results or inference results. Defaults to inference results." } , ) A: Optional[str] = field( default=__lowerCAmelCase , metadata={"help": "Filename under which the plot will be saved. If unused no plot is saved."} , ) A: Optional[List[str]] = list_field( default=__lowerCAmelCase , metadata={"help": "List of model names that are used instead of the ones in the csv file."}) def _a ( SCREAMING_SNAKE_CASE : Any ): """simple docstring""" try: int(SCREAMING_SNAKE_CASE ) return True except ValueError: return False def _a ( SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" try: float(SCREAMING_SNAKE_CASE ) return True except ValueError: return False class __magic_name__ : def __init__( self : Any , lowerCamelCase__ : Dict ) -> Dict: '''simple docstring''' UpperCamelCase__ : int = args UpperCamelCase__ : Any = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='''''' ) as csv_file: UpperCamelCase__ : Union[str, Any] = csv.DictReader(lowerCamelCase__ ) for row in reader: UpperCamelCase__ : Union[str, Any] = row['''model'''] self.result_dict[model_name]["bsz"].append(int(row['''batch_size'''] ) ) self.result_dict[model_name]["seq_len"].append(int(row['''sequence_length'''] ) ) if can_convert_to_int(row['''result'''] ): # value is not None UpperCamelCase__ : Any = int(row['''result'''] ) elif can_convert_to_float(row['''result'''] ): # value is not None UpperCamelCase__ : Any = float(row['''result'''] ) def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ : str = plt.subplots() UpperCamelCase__ : Dict = '''Time usage''' if self.args.is_time else '''Memory usage''' UpperCamelCase__ : int = title_str + ''' for training''' if self.args.is_train else title_str + ''' for inference''' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('''log''' ) ax.set_yscale('''log''' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): UpperCamelCase__ : Tuple = sorted(set(self.result_dict[model_name]['''bsz'''] ) ) UpperCamelCase__ : Tuple = sorted(set(self.result_dict[model_name]['''seq_len'''] ) ) UpperCamelCase__ : Dict = self.result_dict[model_name]['''result'''] ((UpperCamelCase__) , (UpperCamelCase__)) : int = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) UpperCamelCase__ : Optional[int] = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: UpperCamelCase__ : Optional[Any] = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=lowerCamelCase__ , ) else: UpperCamelCase__ : Tuple = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((UpperCamelCase__) , (UpperCamelCase__)) : str = ( ('''batch_size''', '''len''') if self.args.plot_along_batch else ('''in #tokens''', '''bsz''') ) UpperCamelCase__ : Optional[Any] = np.asarray(lowerCamelCase__ , lowerCamelCase__ )[: len(lowerCamelCase__ )] plt.scatter( lowerCamelCase__ , lowerCamelCase__ , label=F"{label_model_name} - {inner_loop_label}: {inner_loop_value}" ) plt.plot(lowerCamelCase__ , lowerCamelCase__ , '''--''' ) title_str += F" {label_model_name} vs." UpperCamelCase__ : Optional[Any] = title_str[:-4] UpperCamelCase__ : List[Any] = '''Time in s''' if self.args.is_time else '''Memory in MB''' # plot plt.title(lowerCamelCase__ ) plt.xlabel(lowerCamelCase__ ) plt.ylabel(lowerCamelCase__ ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def _a ( ): """simple docstring""" UpperCamelCase__ : Optional[Any] = HfArgumentParser(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = parser.parse_args_into_dataclasses()[0] UpperCamelCase__ : Dict = Plot(args=SCREAMING_SNAKE_CASE ) plot.plot() if __name__ == "__main__": main()
51
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class lowerCAmelCase__ ( metaclass=lowerCamelCase_ ): lowerCAmelCase_ = ['''sentencepiece'''] def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] )
93
from abc import ABC, abstractmethod from argparse import ArgumentParser class snake_case_ (lowerCamelCase_ ): @staticmethod @abstractmethod def lowerCamelCase__( __snake_case :ArgumentParser ) -> Dict: raise NotImplementedError() @abstractmethod def lowerCamelCase__( self :Union[str, Any] ) -> Dict: raise NotImplementedError()
240
0
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets lowerCamelCase : Dict ='''\ @inproceedings{popovic-2015-chrf, title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation", month = sep, year = "2015", address = "Lisbon, Portugal", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W15-3049", doi = "10.18653/v1/W15-3049", pages = "392--395", } @inproceedings{popovic-2017-chrf, title = "chr{F}++: words helping character n-grams", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Second Conference on Machine Translation", month = sep, year = "2017", address = "Copenhagen, Denmark", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W17-4770", doi = "10.18653/v1/W17-4770", pages = "612--618", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' lowerCamelCase : Optional[int] ='''\ ChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches, and ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation that is already present in sacrebleu. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information. ''' lowerCamelCase : str =''' Produces ChrF(++) scores for hypotheses given reference translations. Args: predictions (list of str): The predicted sentences. references (list of list of str): The references. There should be one reference sub-list for each prediction sentence. char_order (int): Character n-gram order. Defaults to `6`. word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`. beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`. lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`. whitespace (bool): If `True`, include whitespaces when extracting character n-grams. eps_smoothing (bool): If `True`, applies epsilon smoothing similar to reference chrF++.py, NLTK and Moses implementations. If `False`, it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`. Returns: \'score\' (float): The chrF (chrF++) score, \'char_order\' (int): The character n-gram order, \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++, \'beta\' (int): Determine the importance of recall w.r.t precision Examples: Example 1--a simple example of calculating chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, references=reference) >>> print(results) {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2} Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2) >>> print(results) {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2, ... lowercase=True) >>> print(results) {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __a ( datasets.Metric ): def __lowercase ( self : Tuple ): '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/mjpost/sacreBLEU#chrf--chrf" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#chrf--chrf"] , reference_urls=[ "https://github.com/m-popovic/chrF", ] , ) def __lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : int = CHRF.CHAR_ORDER , SCREAMING_SNAKE_CASE : int = CHRF.WORD_ORDER , SCREAMING_SNAKE_CASE : int = CHRF.BETA , SCREAMING_SNAKE_CASE : bool = False , SCREAMING_SNAKE_CASE : bool = False , SCREAMING_SNAKE_CASE : bool = False , ): '''simple docstring''' UpperCamelCase__ : List[str] = len(references[0] ) if any(len(SCREAMING_SNAKE_CASE ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) UpperCamelCase__ : Union[str, Any] = [[refs[i] for refs in references] for i in range(SCREAMING_SNAKE_CASE )] UpperCamelCase__ : Optional[Any] = CHRF(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Tuple = sb_chrf.corpus_score(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
196
lowerCamelCase : Optional[int] ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> list[str]: UpperCamelCase__ : Optional[Any] = set() # keep track of all the paths to be checked UpperCamelCase__ : Optional[Any] = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue UpperCamelCase__ : int = queue.pop(0 ) # get the last node from the path UpperCamelCase__ : Dict = path[-1] if node not in explored: UpperCamelCase__ : Tuple = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: UpperCamelCase__ : List[str] = list(__lowerCAmelCase ) new_path.append(__lowerCAmelCase ) queue.append(__lowerCAmelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(__lowerCAmelCase ) # in case there's no path between the 2 nodes return [] def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> int: if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 UpperCamelCase__ : Tuple = [start] UpperCamelCase__ : Optional[int] = set(__lowerCAmelCase ) # Keep tab on distances from `start` node. UpperCamelCase__ : str = {start: 0, target: -1} while queue: UpperCamelCase__ : Any = queue.pop(0 ) if node == target: UpperCamelCase__ : Union[str, Any] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(__lowerCAmelCase ) queue.append(__lowerCAmelCase ) UpperCamelCase__ : List[Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, '''G''', '''D''')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, '''G''', '''D''')) # returns 4
196
1
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __lowercase : List[Any] = logging.get_logger(__name__) class __lowercase ( _lowercase ): lowerCamelCase : List[str] = ["input_features"] def __init__(self , A=8_0 , A=1_6_0_0_0 , A=1_6_0 , A=3_0 , A=4_0_0 , A=0.0 , A=False , **A , ): super().__init__( feature_size=A , sampling_rate=A , padding_value=A , return_attention_mask=A , **A , ) lowerCamelCase_ : Dict = n_fft lowerCamelCase_ : Tuple = hop_length lowerCamelCase_ : str = chunk_length lowerCamelCase_ : Optional[Any] = chunk_length * sampling_rate lowerCamelCase_ : int = self.n_samples // hop_length lowerCamelCase_ : Dict = sampling_rate lowerCamelCase_ : Optional[Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A , min_frequency=0.0 , max_frequency=80_00.0 , sampling_rate=A , norm='''slaney''' , mel_scale='''slaney''' , ) def UpperCAmelCase__ (self , A ): lowerCamelCase_ : Tuple = spectrogram( A , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) lowerCamelCase_ : str = log_spec[:, :-1] lowerCamelCase_ : List[str] = np.maximum(A , log_spec.max() - 8.0 ) lowerCamelCase_ : Dict = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCAmelCase__ (A , A , A = 0.0 ): if attention_mask is not None: lowerCamelCase_ : Dict = np.array(A , np.intaa ) lowerCamelCase_ : Dict = [] for vector, length in zip(A , attention_mask.sum(-1 ) ): lowerCamelCase_ : List[Any] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: lowerCamelCase_ : Optional[int] = padding_value normed_input_values.append(A ) else: lowerCamelCase_ : Dict = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__(self , A , A = True , A = None , A = None , A = None , A = "max_length" , A = None , A = None , A = None , **A , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"""The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a""" F""" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input""" F""" was sampled with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) lowerCamelCase_ : Optional[Any] = isinstance(A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) lowerCamelCase_ : Optional[int] = is_batched_numpy or ( isinstance(A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCamelCase_ : List[Any] = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A , np.ndarray ): lowerCamelCase_ : int = np.asarray(A , dtype=np.floataa ) elif isinstance(A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCamelCase_ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCamelCase_ : Optional[Any] = [np.asarray([raw_speech] ).T] lowerCamelCase_ : str = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding lowerCamelCase_ : Any = self.pad( A , padding=A , max_length=max_length if max_length else self.n_samples , truncation=A , pad_to_multiple_of=A , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: lowerCamelCase_ : List[Any] = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) lowerCamelCase_ : Dict = np.stack(padded_inputs['''input_features'''] , axis=0 ) # make sure list is in array format lowerCamelCase_ : int = padded_inputs.get('''input_features''' ).transpose(2 , 0 , 1 ) lowerCamelCase_ : Any = [self._np_extract_fbank_features(A ) for waveform in input_features[0]] if isinstance(input_features[0] , A ): lowerCamelCase_ : List[str] = [np.asarray(A , dtype=np.floataa ) for feature in input_features] else: lowerCamelCase_ : Union[str, Any] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) lowerCamelCase_ : List[Any] = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: lowerCamelCase_ : int = padded_inputs.convert_to_tensors(A ) return padded_inputs def UpperCAmelCase__ (self ): lowerCamelCase_ : Optional[Any] = copy.deepcopy(self.__dict__ ) lowerCamelCase_ : Any = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
318
'''simple docstring''' import numpy as np def lowercase_ ( _lowercase ) -> np.ndarray: '''simple docstring''' return 1 / (1 + np.exp(-vector )) def lowercase_ ( _lowercase ) -> np.ndarray: '''simple docstring''' return vector * sigmoid(_lowercase ) if __name__ == "__main__": import doctest doctest.testmod()
318
1
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def UpperCAmelCase_ (_lowerCAmelCase : int ): # A local function to see if a dot lands in the circle. def is_in_circle(_lowerCAmelCase : float , _lowerCAmelCase : float ) -> bool: __UpperCamelCase : int = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle __UpperCamelCase : str = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(_lowerCAmelCase ) ) # The ratio of the area for circle to square is pi/4. __UpperCamelCase : Optional[int] = proportion * 4 print(F'''The estimated value of pi is {pi_estimate}''' ) print(F'''The numpy value of pi is {pi}''' ) print(F'''The total error is {abs(pi - pi_estimate )}''' ) def UpperCAmelCase_ (_lowerCAmelCase : int , _lowerCAmelCase : Callable[[float], float] , _lowerCAmelCase : float = 0.0 , _lowerCAmelCase : float = 1.0 , ): return mean( function_to_integrate(uniform(_lowerCAmelCase , _lowerCAmelCase ) ) for _ in range(_lowerCAmelCase ) ) * (max_value - min_value) def UpperCAmelCase_ (_lowerCAmelCase : int , _lowerCAmelCase : float = 0.0 , _lowerCAmelCase : float = 1.0 ): def identity_function(_lowerCAmelCase : float ) -> float: return x __UpperCamelCase : str = area_under_curve_estimator( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : Union[str, Any] = (max_value * max_value - min_value * min_value) / 2 print("******************" ) print(F'''Estimating area under y=x where x varies from {min_value} to {max_value}''' ) print(F'''Estimated value is {estimated_value}''' ) print(F'''Expected value is {expected_value}''' ) print(F'''Total error is {abs(estimated_value - expected_value )}''' ) print("******************" ) def UpperCAmelCase_ (_lowerCAmelCase : int ): def function_to_integrate(_lowerCAmelCase : float ) -> float: return sqrt(4.0 - x * x ) __UpperCamelCase : str = area_under_curve_estimator( _lowerCAmelCase , _lowerCAmelCase , 0.0 , 2.0 ) print("******************" ) print("Estimating pi using area_under_curve_estimator" ) print(F'''Estimated value is {estimated_value}''' ) print(F'''Expected value is {pi}''' ) print(F'''Total error is {abs(estimated_value - pi )}''' ) print("******************" ) if __name__ == "__main__": import doctest doctest.testmod()
171
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=[10, 20, 30, 40] , __UpperCamelCase=[2, 2, 3, 2] , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=10 , __UpperCamelCase=0.02 , __UpperCamelCase=["stage2", "stage3", "stage4"] , __UpperCamelCase=3 , __UpperCamelCase=None , ) -> str: '''simple docstring''' __UpperCamelCase : Union[str, Any] = parent __UpperCamelCase : List[Any] = batch_size __UpperCamelCase : Union[str, Any] = image_size __UpperCamelCase : Any = num_channels __UpperCamelCase : Union[str, Any] = num_stages __UpperCamelCase : List[Any] = hidden_sizes __UpperCamelCase : Optional[Any] = depths __UpperCamelCase : Dict = is_training __UpperCamelCase : List[Any] = use_labels __UpperCamelCase : str = intermediate_size __UpperCamelCase : int = hidden_act __UpperCamelCase : Tuple = type_sequence_label_size __UpperCamelCase : List[Any] = initializer_range __UpperCamelCase : List[Any] = out_features __UpperCamelCase : Optional[Any] = num_labels __UpperCamelCase : Optional[Any] = scope __UpperCamelCase : List[Any] = num_stages def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCamelCase : Optional[Any] = None if self.use_labels: __UpperCamelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCamelCase : int = self.get_config() return config, pixel_values, labels def __lowerCamelCase ( self ) -> List[str]: '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def __lowerCamelCase ( self ) -> str: '''simple docstring''' return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=__UpperCamelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=__UpperCamelCase , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> List[Any]: '''simple docstring''' __UpperCamelCase : Union[str, Any] = UperNetForSemanticSegmentation(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() __UpperCamelCase : List[str] = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase : int = self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) : Optional[int] = config_and_inputs __UpperCamelCase : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" lowercase : Any = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowercase : Dict = {'image-segmentation': UperNetForSemanticSegmentation} if is_torch_available() else {} lowercase : Union[str, Any] = False lowercase : Tuple = False lowercase : Optional[int] = False lowercase : Tuple = False lowercase : List[str] = False lowercase : Any = False def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase : Tuple = UperNetModelTester(self ) __UpperCamelCase : str = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCamelCase ( self ) -> Tuple: '''simple docstring''' return def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' __UpperCamelCase , __UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase : Tuple = model_class(__UpperCamelCase ) __UpperCamelCase : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCamelCase : List[str] = [*signature.parameters.keys()] __UpperCamelCase : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @unittest.skip(reason="UperNet does not use inputs_embeds" ) def __lowerCamelCase ( self ) -> List[Any]: '''simple docstring''' pass @unittest.skip(reason="UperNet does not support input and output embeddings" ) def __lowerCamelCase ( self ) -> List[str]: '''simple docstring''' pass @unittest.skip(reason="UperNet does not have a base model" ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip(reason="UperNet does not have a base model" ) def __lowerCamelCase ( self ) -> List[Any]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason="UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' pass def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): __UpperCamelCase : int = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): __UpperCamelCase : Any = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) __UpperCamelCase : Optional[int] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCamelCase : int = self.model_tester.num_stages self.assertEqual(len(__UpperCamelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __UpperCamelCase , __UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase : Any = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCamelCase : List[str] = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase , __UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase : Tuple = _config_zero_init(__UpperCamelCase ) __UpperCamelCase : int = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: __UpperCamelCase : List[str] = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip(reason="UperNet does not have tied weights" ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' pass @slow def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : str = UperNetForSemanticSegmentation.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def UpperCAmelCase_ (): __UpperCamelCase : Union[str, Any] = hf_hub_download( repo_id="hf-internal-testing/fixtures_ade20k" , repo_type="dataset" , filename="ADE_val_00000001.jpg" ) __UpperCamelCase : List[str] = Image.open(_lowerCAmelCase ).convert("RGB" ) return image @require_torch @require_vision @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def __lowerCamelCase ( self ) -> str: '''simple docstring''' __UpperCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained("openmmlab/upernet-swin-tiny" ) __UpperCamelCase : Dict = UperNetForSemanticSegmentation.from_pretrained("openmmlab/upernet-swin-tiny" ).to(__UpperCamelCase ) __UpperCamelCase : Dict = prepare_img() __UpperCamelCase : Any = processor(images=__UpperCamelCase , return_tensors="pt" ).to(__UpperCamelCase ) with torch.no_grad(): __UpperCamelCase : Any = model(**__UpperCamelCase ) __UpperCamelCase : Tuple = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) __UpperCamelCase : Union[str, Any] = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained("openmmlab/upernet-convnext-tiny" ) __UpperCamelCase : List[Any] = UperNetForSemanticSegmentation.from_pretrained("openmmlab/upernet-convnext-tiny" ).to(__UpperCamelCase ) __UpperCamelCase : Dict = prepare_img() __UpperCamelCase : int = processor(images=__UpperCamelCase , return_tensors="pt" ).to(__UpperCamelCase ) with torch.no_grad(): __UpperCamelCase : int = model(**__UpperCamelCase ) __UpperCamelCase : Dict = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) __UpperCamelCase : Union[str, Any] = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
171
1
from __future__ import annotations __lowerCamelCase : Tuple = '''Muhammad Umer Farooq''' __lowerCamelCase : int = '''MIT''' __lowerCamelCase : Union[str, Any] = '''1.0.0''' __lowerCamelCase : List[Any] = '''Muhammad Umer Farooq''' __lowerCamelCase : Optional[Any] = '''contact@muhammadumerfarooq.me''' __lowerCamelCase : Dict = '''Alpha''' import re from html.parser import HTMLParser from urllib import parse import requests class a__ ( A__ ): def __init__( self : str,_A : str ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE_ : list[str] = [] SCREAMING_SNAKE_CASE_ : Tuple = domain def __UpperCamelCase ( self : str,_A : str,_A : list[tuple[str, str | None]] ): """simple docstring""" if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: SCREAMING_SNAKE_CASE_ : Tuple = parse.urljoin(self.domain,_A ) self.urls.append(_A ) def _snake_case ( lowerCAmelCase : str ): """simple docstring""" return ".".join(get_sub_domain_name(lowerCAmelCase ).split("." )[-2:] ) def _snake_case ( lowerCAmelCase : str ): """simple docstring""" return parse.urlparse(lowerCAmelCase ).netloc def _snake_case ( lowerCAmelCase : str = "https://github.com" ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = get_domain_name(lowerCAmelCase ) # Initialize the parser SCREAMING_SNAKE_CASE_ : Any = Parser(lowerCAmelCase ) try: # Open URL SCREAMING_SNAKE_CASE_ : int = requests.get(lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through SCREAMING_SNAKE_CASE_ : Tuple = set() for link in parser.urls: # open URL. # read = requests.get(link) try: SCREAMING_SNAKE_CASE_ : List[str] = requests.get(lowerCAmelCase ) # Get the valid email. SCREAMING_SNAKE_CASE_ : int = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Dict = emails_from_url('''https://github.com''') print(f'''{len(emails)} emails found:''') print('''\n'''.join(sorted(emails)))
18
from __future__ import annotations def UpperCamelCase ( __magic_name__ : list[float] , __magic_name__ : list[float] ) -> float: """simple docstring""" lowercase__ = sorted(numsa + numsa ) lowercase__ , lowercase__ = divmod(len(__magic_name__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() A : Any = [float(x) for x in input('Enter the elements of first array: ').split()] A : Union[str, Any] = [float(x) for x in input('Enter the elements of second array: ').split()] print(F'The median of two arrays is: {median_of_two_arrays(array_a, array_a)}')
305
0
def snake_case__ ( lowerCamelCase__ : Optional[int] ) -> Any: if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(_lowerCAmelCase ).count('''1''' ) if __name__ == "__main__": import doctest doctest.testmod()
353
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case__ = logging.get_logger(__name__) snake_case__ = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class UpperCamelCase_ (a__, a__ ): """simple docstring""" _lowerCAmelCase = 'swin' _lowerCAmelCase = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : Any , _lowerCamelCase : Optional[Any]=224 , _lowerCamelCase : List[str]=4 , _lowerCamelCase : Optional[Any]=3 , _lowerCamelCase : Tuple=96 , _lowerCamelCase : List[Any]=[2, 2, 6, 2] , _lowerCamelCase : List[str]=[3, 6, 12, 24] , _lowerCamelCase : List[Any]=7 , _lowerCamelCase : Optional[int]=4.0 , _lowerCamelCase : List[str]=True , _lowerCamelCase : List[str]=0.0 , _lowerCamelCase : Any=0.0 , _lowerCamelCase : Dict=0.1 , _lowerCamelCase : List[str]="gelu" , _lowerCamelCase : Tuple=False , _lowerCamelCase : Dict=0.02 , _lowerCamelCase : Optional[Any]=1E-5 , _lowerCamelCase : Any=32 , _lowerCamelCase : Tuple=None , _lowerCamelCase : Any=None , **_lowerCamelCase : str , ): """simple docstring""" super().__init__(**_lowerCamelCase ) A_ : Optional[int] = image_size A_ : Optional[int] = patch_size A_ : Optional[int] = num_channels A_ : Any = embed_dim A_ : List[Any] = depths A_ : Any = len(_lowerCamelCase ) A_ : List[Any] = num_heads A_ : Tuple = window_size A_ : Tuple = mlp_ratio A_ : Dict = qkv_bias A_ : List[str] = hidden_dropout_prob A_ : List[str] = attention_probs_dropout_prob A_ : Any = drop_path_rate A_ : List[Any] = hidden_act A_ : Tuple = use_absolute_embeddings A_ : int = layer_norm_eps A_ : Optional[Any] = initializer_range A_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model A_ : str = int(embed_dim * 2 ** (len(_lowerCamelCase ) - 1) ) A_ : str = ['''stem'''] + [f'stage{idx}' for idx in range(1 , len(_lowerCamelCase ) + 1 )] A_ ,A_ : Optional[Any] = get_aligned_output_features_output_indices( out_features=_lowerCamelCase , out_indices=_lowerCamelCase , stage_names=self.stage_names ) class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = version.parse('1.11' ) @property def _a ( self : str ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def _a ( self : Union[str, Any] ): """simple docstring""" return 1E-4
4
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __lowerCamelCase : Tuple = None __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) __lowerCamelCase : str = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} __lowerCamelCase : Optional[int] = { '''vocab_file''': { '''google/bigbird-roberta-base''': '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model''', '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model''' ), }, '''tokenizer_file''': { '''google/bigbird-roberta-base''': ( '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json''' ), '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json''' ), }, } __lowerCamelCase : List[str] = { '''google/bigbird-roberta-base''': 40_96, '''google/bigbird-roberta-large''': 40_96, '''google/bigbird-base-trivia-itc''': 40_96, } __lowerCamelCase : Optional[Any] = '''▁''' class a__ ( A__ ): A = VOCAB_FILES_NAMES A = PRETRAINED_VOCAB_FILES_MAP A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A = BigBirdTokenizer A = ['input_ids', 'attention_mask'] A = [] def __init__( self : Union[str, Any],_A : Any=None,_A : Any=None,_A : str="<unk>",_A : str="<s>",_A : int="</s>",_A : Union[str, Any]="<pad>",_A : Dict="[SEP]",_A : int="[MASK]",_A : int="[CLS]",**_A : Any,): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else bos_token SCREAMING_SNAKE_CASE_ : int = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else eos_token SCREAMING_SNAKE_CASE_ : Dict = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else unk_token SCREAMING_SNAKE_CASE_ : str = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else pad_token SCREAMING_SNAKE_CASE_ : int = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else cls_token SCREAMING_SNAKE_CASE_ : List[str] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else sep_token # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE_ : List[Any] = AddedToken(_A,lstrip=_A,rstrip=_A ) if isinstance(_A,_A ) else mask_token super().__init__( _A,tokenizer_file=_A,bos_token=_A,eos_token=_A,unk_token=_A,sep_token=_A,pad_token=_A,cls_token=_A,mask_token=_A,**_A,) SCREAMING_SNAKE_CASE_ : Optional[int] = vocab_file SCREAMING_SNAKE_CASE_ : Union[str, Any] = False if not self.vocab_file else True def __UpperCamelCase ( self : Union[str, Any],_A : List[int],_A : Optional[List[int]] = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __UpperCamelCase ( self : Union[str, Any],_A : List[int],_A : Optional[List[int]] = None,_A : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(_A )) + [1] return [1] + ([0] * len(_A )) + [1] + ([0] * len(_A )) + [1] def __UpperCamelCase ( self : List[Any],_A : List[int],_A : Optional[List[int]] = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : str,_A : str,_A : Optional[str] = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(_A ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return SCREAMING_SNAKE_CASE_ : List[str] = os.path.join( _A,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_A ): copyfile(self.vocab_file,_A ) return (out_vocab_file,)
18
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : str = {} class __snake_case ( a ): UpperCAmelCase__ : str = '''llama''' UpperCAmelCase__ : Dict = ['''past_key_values'''] def __init__( self : str , _snake_case : List[str]=32000 , _snake_case : int=4096 , _snake_case : List[str]=11008 , _snake_case : Optional[int]=32 , _snake_case : List[Any]=32 , _snake_case : Tuple=None , _snake_case : int="silu" , _snake_case : List[Any]=2048 , _snake_case : List[str]=0.0_2 , _snake_case : Any=1e-6 , _snake_case : List[str]=True , _snake_case : Optional[Any]=0 , _snake_case : Dict=1 , _snake_case : List[Any]=2 , _snake_case : str=1 , _snake_case : Union[str, Any]=False , _snake_case : str=None , **_snake_case : List[Any] , ): """simple docstring""" UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads # for backward compatibility if num_key_value_heads is None: UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = num_key_value_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = initializer_range UpperCAmelCase_ = rms_norm_eps UpperCAmelCase_ = pretraining_tp UpperCAmelCase_ = use_cache UpperCAmelCase_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , tie_word_embeddings=_snake_case , **_snake_case , ) def lowerCamelCase ( self : Optional[int]): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _snake_case) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""") UpperCAmelCase_ = self.rope_scaling.get('''type''' , _snake_case) UpperCAmelCase_ = self.rope_scaling.get('''factor''' , _snake_case) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""") if rope_scaling_factor is None or not isinstance(_snake_case , _snake_case) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""")
51
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging A : Dict = logging.get_logger(__name__) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = ['''pixel_values'''] def __init__( self : Union[str, Any] , __magic_name__ : bool = True , __magic_name__ : Optional[Dict[str, int]] = None , __magic_name__ : PILImageResampling = PILImageResampling.BILINEAR , __magic_name__ : bool = True , __magic_name__ : Dict[str, int] = None , __magic_name__ : bool = True , __magic_name__ : Union[int, float] = 1 / 255 , __magic_name__ : bool = True , __magic_name__ : Optional[Union[float, List[float]]] = None , __magic_name__ : Optional[Union[float, List[float]]] = None , **__magic_name__ : List[str] , ) -> None: super().__init__(**__magic_name__ ) SCREAMING_SNAKE_CASE_ = size if size is not None else {"shortest_edge": 256} SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) SCREAMING_SNAKE_CASE_ = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ ) SCREAMING_SNAKE_CASE_ = do_resize SCREAMING_SNAKE_CASE_ = size SCREAMING_SNAKE_CASE_ = resample SCREAMING_SNAKE_CASE_ = do_center_crop SCREAMING_SNAKE_CASE_ = crop_size SCREAMING_SNAKE_CASE_ = do_rescale SCREAMING_SNAKE_CASE_ = rescale_factor SCREAMING_SNAKE_CASE_ = do_normalize SCREAMING_SNAKE_CASE_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def __A ( self : List[Any] , __magic_name__ : np.ndarray , __magic_name__ : Dict[str, int] , __magic_name__ : PILImageResampling = PILImageResampling.BICUBIC , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : str , ) -> np.ndarray: SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) SCREAMING_SNAKE_CASE_ = get_resize_output_image_size(__magic_name__ , size=size["shortest_edge"] , default_to_square=__magic_name__ ) return resize(__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def __A ( self : List[Any] , __magic_name__ : np.ndarray , __magic_name__ : Dict[str, int] , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : Any , ) -> np.ndarray: SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ ) return center_crop(__magic_name__ , size=(size["height"], size["width"]) , data_format=__magic_name__ , **__magic_name__ ) def __A ( self : Union[str, Any] , __magic_name__ : np.ndarray , __magic_name__ : float , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : Union[str, Any] ) -> np.ndarray: return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def __A ( self : Union[str, Any] , __magic_name__ : np.ndarray , __magic_name__ : Union[float, List[float]] , __magic_name__ : Union[float, List[float]] , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : str , ) -> np.ndarray: return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def __A ( self : str , __magic_name__ : ImageInput , __magic_name__ : Optional[bool] = None , __magic_name__ : Dict[str, int] = None , __magic_name__ : PILImageResampling = None , __magic_name__ : bool = None , __magic_name__ : Dict[str, int] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[float] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[Union[float, List[float]]] = None , __magic_name__ : Optional[Union[float, List[float]]] = None , __magic_name__ : Optional[Union[str, TensorType]] = None , __magic_name__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **__magic_name__ : Dict , ) -> Any: SCREAMING_SNAKE_CASE_ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE_ = size if size is not None else self.size SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) SCREAMING_SNAKE_CASE_ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE_ = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE_ = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE_ = get_size_dict(__magic_name__ ) SCREAMING_SNAKE_CASE_ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE_ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE_ = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE_ = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE_ = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE_ = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE_ = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: SCREAMING_SNAKE_CASE_ = [self.resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE_ = [self.center_crop(image=__magic_name__ , size=__magic_name__ ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE_ = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE_ = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] SCREAMING_SNAKE_CASE_ = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] SCREAMING_SNAKE_CASE_ = {"pixel_values": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
305
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging A : List[str] = logging.get_logger(__name__) A : int = { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = '''blenderbot-small''' lowerCamelCase__ = ['''past_key_values'''] lowerCamelCase__ = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Dict , __magic_name__ : Dict=50_265 , __magic_name__ : str=512 , __magic_name__ : List[Any]=8 , __magic_name__ : Any=2_048 , __magic_name__ : Dict=16 , __magic_name__ : Any=8 , __magic_name__ : Optional[int]=2_048 , __magic_name__ : Dict=16 , __magic_name__ : Tuple=0.0 , __magic_name__ : Dict=0.0 , __magic_name__ : Optional[int]=True , __magic_name__ : Any=True , __magic_name__ : Dict="gelu" , __magic_name__ : Tuple=512 , __magic_name__ : List[str]=0.1 , __magic_name__ : List[Any]=0.0 , __magic_name__ : List[Any]=0.0 , __magic_name__ : Tuple=0.02 , __magic_name__ : Union[str, Any]=1 , __magic_name__ : List[Any]=False , __magic_name__ : str=0 , __magic_name__ : Dict=1 , __magic_name__ : str=2 , __magic_name__ : Union[str, Any]=2 , **__magic_name__ : Optional[Any] , ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = d_model SCREAMING_SNAKE_CASE_ = encoder_ffn_dim SCREAMING_SNAKE_CASE_ = encoder_layers SCREAMING_SNAKE_CASE_ = encoder_attention_heads SCREAMING_SNAKE_CASE_ = decoder_ffn_dim SCREAMING_SNAKE_CASE_ = decoder_layers SCREAMING_SNAKE_CASE_ = decoder_attention_heads SCREAMING_SNAKE_CASE_ = dropout SCREAMING_SNAKE_CASE_ = attention_dropout SCREAMING_SNAKE_CASE_ = activation_dropout SCREAMING_SNAKE_CASE_ = activation_function SCREAMING_SNAKE_CASE_ = init_std SCREAMING_SNAKE_CASE_ = encoder_layerdrop SCREAMING_SNAKE_CASE_ = decoder_layerdrop SCREAMING_SNAKE_CASE_ = use_cache SCREAMING_SNAKE_CASE_ = encoder_layers SCREAMING_SNAKE_CASE_ = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , is_encoder_decoder=__magic_name__ , decoder_start_token_id=__magic_name__ , forced_eos_token_id=__magic_name__ , **__magic_name__ , ) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def __A ( self : str ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: SCREAMING_SNAKE_CASE_ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: SCREAMING_SNAKE_CASE_ = {0: "batch"} SCREAMING_SNAKE_CASE_ = {0: "batch", 1: "past_decoder_sequence + sequence"} else: SCREAMING_SNAKE_CASE_ = {0: "batch", 1: "decoder_sequence"} SCREAMING_SNAKE_CASE_ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. SCREAMING_SNAKE_CASE_ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_layers for i in range(__magic_name__ ): SCREAMING_SNAKE_CASE_ = {0: "batch", 2: "past_sequence + sequence"} SCREAMING_SNAKE_CASE_ = {0: "batch", 2: "past_sequence + sequence"} else: SCREAMING_SNAKE_CASE_ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def __A ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: SCREAMING_SNAKE_CASE_ = super().outputs else: SCREAMING_SNAKE_CASE_ = super(__magic_name__ , self ).outputs if self.use_past: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_layers for i in range(__magic_name__ ): SCREAMING_SNAKE_CASE_ = {0: "batch", 2: "past_sequence + sequence"} SCREAMING_SNAKE_CASE_ = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def __A ( self : int , __magic_name__ : PreTrainedTokenizer , __magic_name__ : int = -1 , __magic_name__ : int = -1 , __magic_name__ : bool = False , __magic_name__ : Optional[TensorType] = None , ) -> Mapping[str, Any]: SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Generate decoder inputs SCREAMING_SNAKE_CASE_ = seq_length if not self.use_past else 1 SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} SCREAMING_SNAKE_CASE_ = dict(**__magic_name__ , **__magic_name__ ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = common_inputs["input_ids"].shape SCREAMING_SNAKE_CASE_ = common_inputs["decoder_input_ids"].shape[1] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_attention_heads SCREAMING_SNAKE_CASE_ = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) SCREAMING_SNAKE_CASE_ = decoder_seq_length + 3 SCREAMING_SNAKE_CASE_ = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) SCREAMING_SNAKE_CASE_ = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(__magic_name__ , __magic_name__ )] , dim=1 ) SCREAMING_SNAKE_CASE_ = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_layers SCREAMING_SNAKE_CASE_ = min(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = max(__magic_name__ , __magic_name__ ) - min_num_layers SCREAMING_SNAKE_CASE_ = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(__magic_name__ ): common_inputs["past_key_values"].append( ( torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), ) ) # TODO: test this. SCREAMING_SNAKE_CASE_ = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(__magic_name__ , __magic_name__ ): common_inputs["past_key_values"].append((torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) ) return common_inputs def __A ( self : Union[str, Any] , __magic_name__ : PreTrainedTokenizer , __magic_name__ : int = -1 , __magic_name__ : int = -1 , __magic_name__ : bool = False , __magic_name__ : Optional[TensorType] = None , ) -> Mapping[str, Any]: SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = common_inputs["input_ids"].shape # Not using the same length for past_key_values SCREAMING_SNAKE_CASE_ = seqlen + 2 SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_layers SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.num_attention_heads SCREAMING_SNAKE_CASE_ = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) SCREAMING_SNAKE_CASE_ = common_inputs["attention_mask"].dtype SCREAMING_SNAKE_CASE_ = torch.cat( [common_inputs["attention_mask"], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) SCREAMING_SNAKE_CASE_ = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(__magic_name__ ) ] return common_inputs def __A ( self : Dict , __magic_name__ : PreTrainedTokenizer , __magic_name__ : int = -1 , __magic_name__ : int = -1 , __magic_name__ : bool = False , __magic_name__ : Optional[TensorType] = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX SCREAMING_SNAKE_CASE_ = compute_effective_axis_dimension( __magic_name__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX SCREAMING_SNAKE_CASE_ = tokenizer.num_special_tokens_to_add(__magic_name__ ) SCREAMING_SNAKE_CASE_ = compute_effective_axis_dimension( __magic_name__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__magic_name__ ) # Generate dummy inputs according to compute batch and sequence SCREAMING_SNAKE_CASE_ = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size SCREAMING_SNAKE_CASE_ = dict(tokenizer(__magic_name__ , return_tensors=__magic_name__ ) ) return common_inputs def __A ( self : Optional[Any] , __magic_name__ : PreTrainedTokenizer , __magic_name__ : int = -1 , __magic_name__ : int = -1 , __magic_name__ : bool = False , __magic_name__ : Optional[TensorType] = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) elif self.task == "causal-lm": SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_causal_lm( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) else: SCREAMING_SNAKE_CASE_ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) return common_inputs def __A ( self : Optional[Any] , __magic_name__ : str , __magic_name__ : List[Any] , __magic_name__ : str , __magic_name__ : List[str] ) -> List[str]: if self.task in ["default", "seq2seq-lm"]: SCREAMING_SNAKE_CASE_ = super()._flatten_past_key_values_(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: SCREAMING_SNAKE_CASE_ = super(__magic_name__ , self )._flatten_past_key_values_( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ )
305
1
__lowerCAmelCase = [0, 2, 4, 6, 8] __lowerCAmelCase = [1, 3, 5, 7, 9] def snake_case_ ( snake_case , snake_case , snake_case , snake_case ) -> int: if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 lowercase__: List[str] = 0 for digit in range(10 ): lowercase__: List[str] = digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , snake_case , snake_case ) return result lowercase__: str = 0 for digita in range(10 ): lowercase__: Union[str, Any] = digita if (remainder + digita) % 2 == 0: lowercase__: Dict = ODD_DIGITS else: lowercase__: Any = EVEN_DIGITS for digita in other_parity_digits: lowercase__: Optional[int] = digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , snake_case , snake_case , ) return result def snake_case_ ( snake_case = 9 ) -> int: lowercase__: List[str] = 0 for length in range(1 , max_power + 1 ): result += reversible_numbers(snake_case , 0 , [0] * length , snake_case ) return result if __name__ == "__main__": print(F'''{solution() = }''')
196
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging __lowerCAmelCase = logging.get_logger(__name__) def snake_case_ ( snake_case , snake_case ) -> List[str]: lowercase__: List[str] = set() lowercase__: List[Any] = [] def parse_line(snake_case ): for line in fp: if isinstance(snake_case , snake_case ): lowercase__: Optional[Any] = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(snake_case ) > 0: lowercase__: List[str] = '\n'.join(snake_case ) # Only keep the warnings specified in `targets` if any(f': {x}: ' in warning for x in targets ): selected_warnings.add(snake_case ) buffer.clear() continue else: lowercase__: Union[str, Any] = line.strip() buffer.append(snake_case ) if from_gh: for filename in os.listdir(snake_case ): lowercase__: Dict = os.path.join(snake_case , snake_case ) if not os.path.isdir(snake_case ): # read the file if filename != "warnings.txt": continue with open(snake_case ) as fp: parse_line(snake_case ) else: try: with zipfile.ZipFile(snake_case ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case ): # read the file if filename != "warnings.txt": continue with z.open(snake_case ) as fp: parse_line(snake_case ) except Exception: logger.warning( f'{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.' ) return selected_warnings def snake_case_ ( snake_case , snake_case ) -> Any: lowercase__: Optional[Any] = set() lowercase__: int = [os.path.join(snake_case , snake_case ) for p in os.listdir(snake_case ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(snake_case , snake_case ) ) return selected_warnings if __name__ == "__main__": def snake_case_ ( snake_case ) -> str: return values.split(',' ) __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') parser.add_argument( '''--output_dir''', type=str, required=True, help='''Where to store the downloaded artifacts and other result files.''', ) parser.add_argument('''--token''', default=None, type=str, help='''A token that has actions:read permission.''') # optional parameters parser.add_argument( '''--targets''', default='''DeprecationWarning,UserWarning,FutureWarning''', type=list_str, help='''Comma-separated list of target warning(s) which we want to extract.''', ) parser.add_argument( '''--from_gh''', action='''store_true''', help='''If running from a GitHub action workflow and collecting warnings from its artifacts.''', ) __lowerCAmelCase = parser.parse_args() __lowerCAmelCase = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links __lowerCAmelCase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, '''artifacts.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print('''=''' * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts __lowerCAmelCase = extract_warnings(args.output_dir, args.targets) __lowerCAmelCase = sorted(selected_warnings) with open(os.path.join(args.output_dir, '''selected_warnings.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
196
1
'''simple docstring''' import re import string import numpy as np import datasets UpperCamelCase : List[Any] = """ Returns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list. """ UpperCamelCase : int = """ Args: predictions: List of predicted texts. references: List of reference texts. regexes_to_ignore: List, defaults to None. Regex expressions of characters to ignore when calculating the exact matches. Note: these regexes are removed from the input data before the changes based on the options below (e.g. ignore_case, ignore_punctuation, ignore_numbers) are applied. ignore_case: Boolean, defaults to False. If true, turns everything to lowercase so that capitalization differences are ignored. ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. Returns: exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive. Examples: >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 25.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 50.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 75.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True) >>> print(round(results[\"exact_match\"], 1)) 100.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"] >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 33.3 """ UpperCamelCase : Union[str, Any] = """ """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self : List[str]): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence'), 'references': datasets.Value('string' , id='sequence'), }) , reference_urls=[] , ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Any , UpperCAmelCase_ : Union[str, Any]=None , UpperCAmelCase_ : int=False , UpperCAmelCase_ : List[Any]=False , UpperCAmelCase_ : Union[str, Any]=False , ): """simple docstring""" if regexes_to_ignore is not None: for s in regexes_to_ignore: a : Union[str, Any] = np.array([re.sub(UpperCAmelCase_ , '' , UpperCAmelCase_) for x in predictions]) a : str = np.array([re.sub(UpperCAmelCase_ , '' , UpperCAmelCase_) for x in references]) else: a : List[Any] = np.asarray(UpperCAmelCase_) a : int = np.asarray(UpperCAmelCase_) if ignore_case: a : List[Any] = np.char.lower(UpperCAmelCase_) a : Optional[int] = np.char.lower(UpperCAmelCase_) if ignore_punctuation: a : List[str] = string.punctuation.maketrans('' , '' , string.punctuation) a : List[Any] = np.char.translate(UpperCAmelCase_ , table=UpperCAmelCase_) a : Tuple = np.char.translate(UpperCAmelCase_ , table=UpperCAmelCase_) if ignore_numbers: a : List[Any] = string.digits.maketrans('' , '' , string.digits) a : str = np.char.translate(UpperCAmelCase_ , table=UpperCAmelCase_) a : str = np.char.translate(UpperCAmelCase_ , table=UpperCAmelCase_) a : Optional[Any] = predictions == references return {"exact_match": np.mean(UpperCAmelCase_) * 1_0_0}
363
'''simple docstring''' import math from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase : str = logging.get_logger(__name__) UpperCamelCase : List[str] = { """facebook/data2vec-base-960h""": """https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json""", # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class UpperCamelCase ( a_ ): """simple docstring""" A : Optional[int] = "data2vec-audio" def __init__( self : Dict , UpperCAmelCase_ : Optional[int]=3_2 , UpperCAmelCase_ : Union[str, Any]=7_6_8 , UpperCAmelCase_ : Dict=1_2 , UpperCAmelCase_ : str=1_2 , UpperCAmelCase_ : Optional[Any]=3_0_7_2 , UpperCAmelCase_ : Dict="gelu" , UpperCAmelCase_ : Union[str, Any]=0.1 , UpperCAmelCase_ : Optional[int]=0.1 , UpperCAmelCase_ : int=0.1 , UpperCAmelCase_ : Union[str, Any]=0.0 , UpperCAmelCase_ : Union[str, Any]=0.1 , UpperCAmelCase_ : Dict=0.1 , UpperCAmelCase_ : Optional[Any]=0.02 , UpperCAmelCase_ : str=1e-5 , UpperCAmelCase_ : List[Any]="gelu" , UpperCAmelCase_ : Optional[Any]=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) , UpperCAmelCase_ : Dict=(5, 2, 2, 2, 2, 2, 2) , UpperCAmelCase_ : int=(1_0, 3, 3, 3, 3, 2, 2) , UpperCAmelCase_ : List[Any]=False , UpperCAmelCase_ : List[Any]=1_6 , UpperCAmelCase_ : Optional[Any]=1_9 , UpperCAmelCase_ : int=5 , UpperCAmelCase_ : Any=0.05 , UpperCAmelCase_ : Dict=1_0 , UpperCAmelCase_ : Optional[Any]=2 , UpperCAmelCase_ : Optional[int]=0.0 , UpperCAmelCase_ : Tuple=1_0 , UpperCAmelCase_ : int=0 , UpperCAmelCase_ : Any="sum" , UpperCAmelCase_ : Optional[int]=False , UpperCAmelCase_ : Dict=False , UpperCAmelCase_ : Optional[int]=2_5_6 , UpperCAmelCase_ : Any=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 1_5_0_0) , UpperCAmelCase_ : Optional[Any]=(5, 3, 3, 1, 1) , UpperCAmelCase_ : Optional[int]=(1, 2, 3, 1, 1) , UpperCAmelCase_ : int=5_1_2 , UpperCAmelCase_ : str=0 , UpperCAmelCase_ : Tuple=1 , UpperCAmelCase_ : Union[str, Any]=2 , UpperCAmelCase_ : List[str]=False , UpperCAmelCase_ : Any=3 , UpperCAmelCase_ : Dict=2 , UpperCAmelCase_ : int=3 , UpperCAmelCase_ : List[Any]=None , **UpperCAmelCase_ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_) a : List[Any] = hidden_size a : Any = feat_extract_activation a : Any = list(UpperCAmelCase_) a : Optional[int] = list(UpperCAmelCase_) a : Dict = list(UpperCAmelCase_) a : Tuple = conv_bias a : str = num_conv_pos_embeddings a : Dict = num_conv_pos_embedding_groups a : Optional[Any] = conv_pos_kernel_size a : Any = len(self.conv_dim) a : Tuple = num_hidden_layers a : Any = intermediate_size a : Any = hidden_act a : Dict = num_attention_heads a : Dict = hidden_dropout a : Union[str, Any] = attention_dropout a : Dict = activation_dropout a : Optional[int] = feat_proj_dropout a : Tuple = final_dropout a : Union[str, Any] = layerdrop a : Tuple = layer_norm_eps a : Dict = initializer_range a : Tuple = vocab_size a : int = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel)}`.""") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 a : List[str] = mask_time_prob a : int = mask_time_length a : Optional[int] = mask_time_min_masks a : Dict = mask_feature_prob a : List[str] = mask_feature_length a : str = mask_feature_min_masks # ctc loss a : str = ctc_loss_reduction a : Optional[Any] = ctc_zero_infinity # adapter a : List[str] = add_adapter a : Optional[Any] = adapter_kernel_size a : int = adapter_stride a : str = num_adapter_layers a : Optional[Any] = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. a : str = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. a : List[Any] = list(UpperCAmelCase_) a : List[str] = list(UpperCAmelCase_) a : str = list(UpperCAmelCase_) a : Optional[Any] = xvector_output_dim @property def SCREAMING_SNAKE_CASE_ ( self : Dict): """simple docstring""" return math.prod(self.conv_stride)
345
0
"""simple docstring""" import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=24 , _lowerCamelCase=2 , _lowerCamelCase=6 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=1000 , ): """simple docstring""" UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : Tuple = seq_length UpperCAmelCase__ : Any = is_training UpperCAmelCase__ : Any = use_input_mask UpperCAmelCase__ : str = use_token_type_ids UpperCAmelCase__ : Optional[Any] = use_labels UpperCAmelCase__ : Optional[int] = vocab_size UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : Union[str, Any] = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Optional[int] = type_vocab_size UpperCAmelCase__ : int = type_sequence_label_size UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : str = num_labels UpperCAmelCase__ : List[Any] = scope UpperCAmelCase__ : int = range_bbox def _a (self ): """simple docstring""" UpperCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCAmelCase__ : Any = bbox[i, j, 3] UpperCAmelCase__ : int = bbox[i, j, 1] UpperCAmelCase__ : Any = t if bbox[i, j, 2] < bbox[i, j, 0]: UpperCAmelCase__ : Optional[Any] = bbox[i, j, 2] UpperCAmelCase__ : int = bbox[i, j, 0] UpperCAmelCase__ : Optional[int] = t UpperCAmelCase__ : int = None if self.use_input_mask: UpperCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) UpperCAmelCase__ : List[str] = None if self.use_token_type_ids: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : Union[str, Any] = None if self.use_labels: UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ : str = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def _a (self ): """simple docstring""" return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Any = LiltModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Optional[Any] = model(_lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase , bbox=_lowerCamelCase , token_type_ids=_lowerCamelCase ) UpperCAmelCase__ : Tuple = model(_lowerCamelCase , bbox=_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Dict = self.num_labels UpperCAmelCase__ : Tuple = LiltForTokenClassification(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Dict = model( _lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : int = LiltForQuestionAnswering(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Dict = model( _lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , start_positions=_lowerCamelCase , end_positions=_lowerCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : int = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Optional[Any] = config_and_inputs UpperCAmelCase__ : List[str] = { """input_ids""": input_ids, """bbox""": bbox, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': LiltModel, 'question-answering': LiltForQuestionAnswering, 'text-classification': LiltForSequenceClassification, 'token-classification': LiltForTokenClassification, 'zero-shot': LiltForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" return True def _a (self ): """simple docstring""" UpperCAmelCase__ : Dict = LiltModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def _a (self ): """simple docstring""" self.config_tester.run_common_tests() def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ : str = type self.model_tester.create_and_check_model(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCamelCase ) @slow def _a (self ): """simple docstring""" for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : Dict = LiltModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) @require_torch @slow class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def _a (self ): """simple docstring""" UpperCAmelCase__ : Dict = LiltModel.from_pretrained("""SCUT-DLVCLab/lilt-roberta-en-base""" ).to(_lowerCamelCase ) UpperCAmelCase__ : Any = torch.tensor([[1, 2]] , device=_lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=_lowerCamelCase ) # forward pass with torch.no_grad(): UpperCAmelCase__ : List[str] = model(input_ids=_lowerCamelCase , bbox=_lowerCamelCase ) UpperCAmelCase__ : Tuple = torch.Size([1, 2, 768] ) UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[-0.0_653, 0.0_950, -0.0_061], [-0.0_545, 0.0_926, -0.0_324]] , device=_lowerCamelCase , ) self.assertTrue(outputs.last_hidden_state.shape , _lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , _lowerCamelCase , atol=1e-3 ) )
171
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=99 , _lowerCamelCase=13 , _lowerCamelCase=16 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=False , _lowerCamelCase=True , _lowerCamelCase=2 , _lowerCamelCase=32 , _lowerCamelCase=4 , _lowerCamelCase=4 , _lowerCamelCase=30 , _lowerCamelCase=0 , _lowerCamelCase=1 , _lowerCamelCase=2 , _lowerCamelCase=None , ): """simple docstring""" UpperCAmelCase__ : int = parent UpperCAmelCase__ : Optional[Any] = batch_size UpperCAmelCase__ : Union[str, Any] = decoder_seq_length # For common tests UpperCAmelCase__ : int = self.decoder_seq_length UpperCAmelCase__ : Optional[Any] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : List[Any] = use_labels UpperCAmelCase__ : Optional[Any] = vocab_size UpperCAmelCase__ : List[Any] = d_model UpperCAmelCase__ : List[str] = d_model UpperCAmelCase__ : Dict = decoder_layers UpperCAmelCase__ : Any = decoder_layers UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : Any = decoder_attention_heads UpperCAmelCase__ : List[str] = decoder_attention_heads UpperCAmelCase__ : List[str] = eos_token_id UpperCAmelCase__ : int = bos_token_id UpperCAmelCase__ : Optional[int] = pad_token_id UpperCAmelCase__ : Any = decoder_start_token_id UpperCAmelCase__ : Dict = use_cache UpperCAmelCase__ : Optional[Any] = max_position_embeddings UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : Dict = decoder_seq_length UpperCAmelCase__ : str = 2 UpperCAmelCase__ : List[str] = 1 def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) UpperCAmelCase__ : Any = None if self.use_labels: UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) UpperCAmelCase__ : Union[str, Any] = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : str = TrOCRDecoder(config=_lowerCamelCase ).to(_lowerCamelCase ).eval() UpperCAmelCase__ : Any = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass UpperCAmelCase__ : Optional[Any] = model(_lowerCamelCase , use_cache=_lowerCamelCase ) UpperCAmelCase__ : str = model(_lowerCamelCase ) UpperCAmelCase__ : Optional[int] = model(_lowerCamelCase , use_cache=_lowerCamelCase ) self.parent.assertTrue(len(_lowerCamelCase ) == len(_lowerCamelCase ) ) self.parent.assertTrue(len(_lowerCamelCase ) == len(_lowerCamelCase ) + 1 ) UpperCAmelCase__ : List[Any] = outputs["""past_key_values"""] # create hypothetical next token and extent to next_input_ids UpperCAmelCase__ : List[str] = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and UpperCAmelCase__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase__ : int = model(_lowerCamelCase )["""last_hidden_state"""] UpperCAmelCase__ : Dict = model(_lowerCamelCase , past_key_values=_lowerCamelCase )["""last_hidden_state"""] # select random slice UpperCAmelCase__ : Any = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase__ : int = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() UpperCAmelCase__ : Dict = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=1e-3 ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_torch class lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () SCREAMING_SNAKE_CASE = (TrOCRForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = TrOCRStandaloneDecoderModelTester(self , is_training=_lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = ConfigTester(self , config_class=_lowerCamelCase ) def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" self.config_tester.run_common_tests() def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*_lowerCamelCase ) def _a (self ): """simple docstring""" return @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def _a (self ): """simple docstring""" pass
171
1
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () _UpperCamelCase = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). _UpperCamelCase = [0, 25, 50] _UpperCamelCase = [25, 50, 75] _UpperCamelCase = fuzz.membership.trimf(X, abca) _UpperCamelCase = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. _UpperCamelCase = np.ones(75) _UpperCamelCase = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) _UpperCamelCase = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) _UpperCamelCase = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) _UpperCamelCase = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) _UpperCamelCase = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] _UpperCamelCase = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) _UpperCamelCase = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] _UpperCamelCase = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] _UpperCamelCase = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('''Young''') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('''Middle aged''') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('''union''') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('''intersection''') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('''complement_a''') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('''difference a/b''') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('''alg_sum''') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('''alg_product''') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('''bdd_sum''') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('''bdd_difference''') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
361
import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params _UpperCamelCase = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ['''memory_attention''', '''encoder_attn'''], ['''attention''', '''attn'''], ['''/''', '''.'''], ['''.LayerNorm.gamma''', '''_layer_norm.weight'''], ['''.LayerNorm.beta''', '''_layer_norm.bias'''], ['''r.layer_''', '''r.layers.'''], ['''output_proj''', '''out_proj'''], ['''ffn.dense_1.''', '''fc2.'''], ['''ffn.dense.''', '''fc1.'''], ['''ffn_layer_norm''', '''final_layer_norm'''], ['''kernel''', '''weight'''], ['''encoder_layer_norm.''', '''encoder.layer_norm.'''], ['''decoder_layer_norm.''', '''decoder.layer_norm.'''], ['''embeddings.weights''', '''shared.weight'''], ] def UpperCamelCase_( snake_case__: int ) -> str: for pegasus_name, hf_name in PATTERNS: UpperCAmelCase__ = k.replace(snake_case__ , snake_case__ ) return k def UpperCamelCase_( snake_case__: dict , snake_case__: dict ) -> PegasusForConditionalGeneration: UpperCAmelCase__ = DEFAULTS.copy() cfg_kwargs.update(snake_case__ ) UpperCAmelCase__ = PegasusConfig(**snake_case__ ) UpperCAmelCase__ = PegasusForConditionalGeneration(snake_case__ ) UpperCAmelCase__ = torch_model.model.state_dict() UpperCAmelCase__ = {} for k, v in tf_weights.items(): UpperCAmelCase__ = rename_state_dict_key(snake_case__ ) if new_k not in sd: raise ValueError(f"could not find new key {new_k} in state dict. (converted from {k})" ) if "dense" in k or "proj" in new_k: UpperCAmelCase__ = v.T UpperCAmelCase__ = torch.tensor(snake_case__ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, f"{new_k}, {k}, {v.shape}, {sd[new_k].shape}" # make sure embedding.padding_idx is respected UpperCAmelCase__ = torch.zeros_like(mapping['shared.weight'][cfg.pad_token_id + 1] ) UpperCAmelCase__ = mapping['shared.weight'] UpperCAmelCase__ = mapping['shared.weight'] UpperCAmelCase__ = {k: torch.zeros_like(snake_case__ ) for k, v in sd.items() if k.endswith('bias' ) and k not in mapping} mapping.update(**snake_case__ ) UpperCAmelCase__ , UpperCAmelCase__ = torch_model.model.load_state_dict(snake_case__ , strict=snake_case__ ) UpperCAmelCase__ = [ k for k in missing if k not in ['encoder.embed_positions.weight', 'decoder.embed_positions.weight'] ] assert unexpected_missing == [], f"no matches found for the following torch keys {unexpected_missing}" assert extra == [], f"no matches found for the following tf keys {extra}" return torch_model def UpperCamelCase_( snake_case__: int="./ckpt/aeslc/model.ckpt-32000" ) -> Dict: UpperCAmelCase__ = tf.train.list_variables(snake_case__ ) UpperCAmelCase__ = {} UpperCAmelCase__ = ['Adafactor', 'global_step'] for name, shape in tqdm(snake_case__ , desc='converting tf checkpoint to dict' ): UpperCAmelCase__ = any(pat in name for pat in ignore_name ) if skip_key: continue UpperCAmelCase__ = tf.train.load_variable(snake_case__ , snake_case__ ) UpperCAmelCase__ = array return tf_weights def UpperCamelCase_( snake_case__: str , snake_case__: str ) -> Optional[Any]: # save tokenizer first UpperCAmelCase__ = Path(snake_case__ ).parent.name UpperCAmelCase__ = task_specific_params[f"summarization_{dataset}"]['max_position_embeddings'] UpperCAmelCase__ = PegasusTokenizer.from_pretrained('sshleifer/pegasus' , model_max_length=snake_case__ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(snake_case__ ) # convert model UpperCAmelCase__ = get_tf_weights_as_numpy(snake_case__ ) UpperCAmelCase__ = task_specific_params[f"summarization_{dataset}"] if dataset == "large": UpperCAmelCase__ = task_specific_params UpperCAmelCase__ = convert_pegasus(snake_case__ , snake_case__ ) torch_model.save_pretrained(snake_case__ ) UpperCAmelCase__ = torch_model.state_dict() sd.pop('model.decoder.embed_positions.weight' ) sd.pop('model.encoder.embed_positions.weight' ) torch.save(snake_case__ , Path(snake_case__ ) / 'pytorch_model.bin' ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''tf_ckpt_path''', type=str, help='''passed to tf.train.list_variables''') parser.add_argument('''save_dir''', default=None, type=str, help='''Path to the output PyTorch model.''') _UpperCamelCase = parser.parse_args() if args.save_dir is None: _UpperCamelCase = Path(args.tf_ckpt_path).parent.name _UpperCamelCase = os.path.join('''pegasus''', dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
335
0
class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ ) -> List[Any]: lowerCamelCase : int = val lowerCamelCase : Optional[int] = None lowerCamelCase : Any = None def _lowercase ( self , UpperCamelCase__ ) -> List[str]: if self.val: if val < self.val: if self.left is None: lowerCamelCase : Union[str, Any] = Node(UpperCamelCase__ ) else: self.left.insert(UpperCamelCase__ ) elif val > self.val: if self.right is None: lowerCamelCase : Optional[Any] = Node(UpperCamelCase__ ) else: self.right.insert(UpperCamelCase__ ) else: lowerCamelCase : List[str] = val def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: # Recursive traversal if root: inorder(root.left ,_SCREAMING_SNAKE_CASE ) res.append(root.val ) inorder(root.right ,_SCREAMING_SNAKE_CASE ) def A ( _SCREAMING_SNAKE_CASE ) -> Dict: # Build BST if len(_SCREAMING_SNAKE_CASE ) == 0: return arr lowerCamelCase : List[Any] = Node(arr[0] ) for i in range(1 ,len(_SCREAMING_SNAKE_CASE ) ): root.insert(arr[i] ) # Traverse BST in order. lowerCamelCase : Optional[Any] = [] inorder(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
48
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __snake_case ="""\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __snake_case ="""\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __snake_case =""" Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def a_ ( lowerCamelCase : str , lowerCamelCase : Union[str, Any] ): return float((preds == labels).mean() ) def a_ ( lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : str="binary" ): lowerCAmelCase = simple_accuracy(lowerCamelCase , lowerCamelCase ) lowerCAmelCase = float(fa_score(y_true=lowerCamelCase , y_pred=lowerCamelCase , average=lowerCamelCase ) ) return { "accuracy": acc, "f1": fa, } def a_ ( lowerCamelCase : List[Any] , lowerCamelCase : List[Any] ): lowerCAmelCase = {} for id_pred, label in zip(lowerCamelCase , lowerCamelCase ): lowerCAmelCase = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' lowerCAmelCase = id_pred['prediction'] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase = [(pred, label)] lowerCAmelCase , lowerCAmelCase = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase , lowerCAmelCase = zip(*lowerCamelCase ) lowerCAmelCase = fa_score(y_true=lowerCamelCase , y_pred=lowerCamelCase , average='macro' ) fas.append(lowerCamelCase ) lowerCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCamelCase ) ) ems.append(lowerCamelCase ) lowerCAmelCase = float(sum(lowerCamelCase ) / len(lowerCamelCase ) ) lowerCAmelCase = sum(lowerCamelCase ) / len(lowerCamelCase ) lowerCAmelCase = float(fa_score(y_true=lowerCamelCase , y_pred=[id_pred['prediction'] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): def __UpperCAmelCase ( self : List[str] ) -> List[Any]: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='numpy' if not self.config_name == 'record' and not self.config_name == 'multirc' else None , ) def __UpperCAmelCase ( self : Union[str, Any] ) -> str: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "prediction_text": datasets.Value('string' ), }, "references": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "answers": datasets.Sequence(datasets.Value('string' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('int64' ), "paragraph": datasets.Value('int64' ), "question": datasets.Value('int64' ), }, "prediction": datasets.Value('int64' ), }, "references": datasets.Value('int64' ), } else: return { "predictions": datasets.Value('int64' ), "references": datasets.Value('int64' ), } def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any] ) -> Any: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(UpperCAmelCase__ , UpperCAmelCase__ )} elif self.config_name == "cb": return acc_and_fa(UpperCAmelCase__ , UpperCAmelCase__ , fa_avg='macro' ) elif self.config_name == "record": lowerCAmelCase = [ { 'qas': [ {'id': ref['idx']['query'], 'answers': [{'text': ans} for ans in ref['answers']]} for ref in references ] } ] lowerCAmelCase = {pred['idx']['query']: pred['prediction_text'] for pred in predictions} return evaluate_record(UpperCAmelCase__ , UpperCAmelCase__ )[0] elif self.config_name == "multirc": return evaluate_multirc(UpperCAmelCase__ , UpperCAmelCase__ ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(UpperCAmelCase__ , UpperCAmelCase__ )} else: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' )
4
0
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("To use the rich extension, install rich with `pip install rich`")
293
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def A__ ( SCREAMING_SNAKE_CASE__ = 3) -> qiskit.result.counts.Counts: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): raise TypeError("""number of qubits must be a integer.""") if number_of_qubits <= 0: raise ValueError("""number of qubits must be > 0.""") if math.floor(SCREAMING_SNAKE_CASE__) != number_of_qubits: raise ValueError("""number of qubits must be exact integer.""") if number_of_qubits > 10: raise ValueError("""number of qubits too large to simulate(>10).""") __snake_case: int = QuantumRegister(SCREAMING_SNAKE_CASE__ , """qr""") __snake_case: List[str] = ClassicalRegister(SCREAMING_SNAKE_CASE__ , """cr""") __snake_case: Optional[Any] = QuantumCircuit(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) __snake_case: Tuple = number_of_qubits for i in range(SCREAMING_SNAKE_CASE__): quantum_circuit.h(number_of_qubits - i - 1) counter -= 1 for j in range(SCREAMING_SNAKE_CASE__): quantum_circuit.cp(np.pi / 2 ** (counter - j) , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) for k in range(number_of_qubits // 2): quantum_circuit.swap(SCREAMING_SNAKE_CASE__ , number_of_qubits - k - 1) # measure all the qubits quantum_circuit.measure(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) # simulate with 10000 shots __snake_case: Union[str, Any] = Aer.get_backend("""qasm_simulator""") __snake_case: Optional[Any] = execute(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , shots=1_0000) return job.result().get_counts(SCREAMING_SNAKE_CASE__) if __name__ == "__main__": print( f'Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}' )
293
1
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def UpperCamelCase ( __magic_name__ : Union[str, Any] , __magic_name__ : Dict ) -> Any: """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer lowercase__ = flax_key_tuple[:-1] + ("""weight""",) lowercase__ = torch.permute(__magic_name__ , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(__magic_name__ ): # linear layer lowercase__ = flax_key_tuple[:-1] + ("""weight""",) lowercase__ = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: lowercase__ = flax_key_tuple[:-1] + ("""weight""",) return flax_key_tuple, flax_tensor def UpperCamelCase ( __magic_name__ : Tuple , __magic_name__ : Optional[int] , __magic_name__ : Dict ) -> str: """simple docstring""" if "metadata" in layer: lowercase__ = layer.split("""metadata""" ) lowercase__ = """""".join(split_layer[0] )[:-1] lowercase__ = [tuple(("""metadata""" + split_layer[1]).split("""/""" ) )] elif "kvstore" in layer: lowercase__ = layer.split("""kvstore""" ) lowercase__ = """""".join(split_layer[0] )[:-1] lowercase__ = [tuple(("""kvstore""" + split_layer[1]).split("""/""" ) )] else: lowercase__ = layer.split("""/""" ) lowercase__ = """/""".join(split_layer[:-1] ) lowercase__ = (split_layer[-1],) if "kvstore/path" in layer: lowercase__ = f'''{switch_checkpoint_path}/{checkpoint_info[layer]}''' elif "kvstore/driver" in layer: lowercase__ = """file""" else: lowercase__ = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def UpperCamelCase ( __magic_name__ : Any , __magic_name__ : List[str] ) -> List[Any]: """simple docstring""" lowercase__ = rename_keys(__magic_name__ ) lowercase__ = {} for k, v in current_block.items(): lowercase__ = v lowercase__ = new_current_block torch.save(__magic_name__ , __magic_name__ ) def UpperCamelCase ( __magic_name__ : List[str] , __magic_name__ : Optional[int] , __magic_name__ : Any , __magic_name__ : Dict , __magic_name__ : str = WEIGHTS_NAME ) -> int: """simple docstring""" lowercase__ = convert_file_size_to_int(__magic_name__ ) lowercase__ = [] lowercase__ = {} lowercase__ = 0 lowercase__ = 0 os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with gfile.GFile(switch_checkpoint_path + """/checkpoint""" , """rb""" ) as fp: lowercase__ = serialization.msgpack_restore(fp.read() )["""optimizer"""]["""target"""] lowercase__ = flatten_dict(__magic_name__ , sep="""/""" ) lowercase__ = {} for layer in checkpoint_info.keys(): lowercase__ , lowercase__ , lowercase__ = get_key_and_tensorstore_dict( __magic_name__ , __magic_name__ , __magic_name__ ) if curr_real_layer_name in all_layers: lowercase__ = content else: lowercase__ = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file lowercase__ = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() lowercase__ = torch.tensor(__magic_name__ ) lowercase__ = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts lowercase__ , lowercase__ = rename_base_flax_keys(tuple(key.split("""/""" ) ) , __magic_name__ ) lowercase__ = """/""".join(__magic_name__ ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: lowercase__ = os.path.join( __magic_name__ , weights_name.replace(""".bin""" , f'''-{len(__magic_name__ )+1:05d}-of-???.bin''' ) ) rename_and_save_block(__magic_name__ , __magic_name__ ) sharded_state_dicts.append(current_block.keys() ) del current_block lowercase__ = {} lowercase__ = 0 lowercase__ = raw_weights.to(getattr(__magic_name__ , __magic_name__ ) ) current_block_size += weight_size total_size += weight_size # Add the last block lowercase__ = os.path.join(__magic_name__ , weights_name.replace(""".bin""" , f'''-{len(__magic_name__ )+1:05d}-of-???.bin''' ) ) rename_and_save_block(__magic_name__ , __magic_name__ ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(__magic_name__ ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index lowercase__ = {} lowercase__ = {} for idx, shard in enumerate(__magic_name__ ): lowercase__ = weights_name.replace( """.bin""" , f'''-{idx+1:05d}-of-{len(__magic_name__ ):05d}.bin''' ) # len(sharded_state_dicts):05d} lowercase__ = os.path.join(__magic_name__ , weights_name.replace(""".bin""" , f'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(__magic_name__ , os.path.join(__magic_name__ , __magic_name__ ) ) lowercase__ = shard for key in shard: lowercase__ = shard_file # Add the metadata lowercase__ = {"""total_size""": total_size} lowercase__ = {"""metadata""": metadata, """weight_map""": weight_map} with open(os.path.join(__magic_name__ , __magic_name__ ) , """w""" , encoding="""utf-8""" ) as f: lowercase__ = json.dumps(__magic_name__ , indent=2 , sort_keys=__magic_name__ ) + """\n""" f.write(__magic_name__ ) return metadata, index if __name__ == "__main__": A : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600', type=str, required=False, help='Path to a directory containing a folder per layer. Follows the original Google format.', ) parser.add_argument('--max_shard_size', default='10GB', required=False, help='Max shard size') parser.add_argument('--dtype', default='bfloat16', type=str, required=False, help='dtype of the saved model') parser.add_argument( '--pytorch_dump_folder_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted', type=str, required=False, help='Path to the output pytorch model.', ) A : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def UpperCamelCase ( ) -> Dict: """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer lowercase__ = SwitchTransformersConfig.from_pretrained("""google/switch-base-8""" ) config.save_pretrained("""/home/arthur_huggingface_co/transformers/switch_converted""" ) lowercase__ = SwitchTransformersForConditionalGeneration.from_pretrained( """/home/arthur_huggingface_co/transformers/switch_converted""" , device_map="""auto""" ) lowercase__ = TaTokenizer.from_pretrained("""t5-small""" ) lowercase__ = """A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""" lowercase__ = tokenizer(__magic_name__ , return_tensors="""pt""" ).input_ids lowercase__ = model.generate(__magic_name__ , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
305
import requests from bsa import BeautifulSoup def UpperCamelCase ( __magic_name__ : str = "AAPL" ) -> str: """simple docstring""" lowercase__ = f'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' lowercase__ = BeautifulSoup(requests.get(__magic_name__ ).text , """html.parser""" ) lowercase__ = """My(6px) Pos(r) smartphone_Mt(6px)""" return soup.find("""div""" , class_=class_ ).find("""span""" ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(F'Current {symbol:<4} stock price is {stock_price(symbol):>8}')
305
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''RWKV/rwkv-4-169m-pile''': '''https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json''', '''RWKV/rwkv-4-430m-pile''': '''https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json''', '''RWKV/rwkv-4-1b5-pile''': '''https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json''', '''RWKV/rwkv-4-3b-pile''': '''https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json''', '''RWKV/rwkv-4-7b-pile''': '''https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json''', '''RWKV/rwkv-4-14b-pile''': '''https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json''', '''RWKV/rwkv-raven-1b5''': '''https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json''', '''RWKV/rwkv-raven-3b''': '''https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json''', '''RWKV/rwkv-raven-7b''': '''https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json''', '''RWKV/rwkv-raven-14b''': '''https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json''', } class lowercase_ ( __SCREAMING_SNAKE_CASE ): A__ : Any = 'rwkv' A__ : Optional[Any] = {'max_position_embeddings': 'context_length'} def __init__( self , __UpperCamelCase=5_0_2_7_7 , __UpperCamelCase=1_0_2_4 , __UpperCamelCase=4_0_9_6 , __UpperCamelCase=3_2 , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=1e-5 , __UpperCamelCase=0 , __UpperCamelCase=0 , __UpperCamelCase=6 , __UpperCamelCase=False , __UpperCamelCase=True , **__UpperCamelCase , ): """simple docstring""" UpperCamelCase_ = vocab_size UpperCamelCase_ = context_length UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = attention_hidden_size if attention_hidden_size is not None else hidden_size UpperCamelCase_ = intermediate_size if intermediate_size is not None else 4 * hidden_size UpperCamelCase_ = layer_norm_epsilon UpperCamelCase_ = rescale_every UpperCamelCase_ = use_cache UpperCamelCase_ = bos_token_id UpperCamelCase_ = eos_token_id super().__init__( tie_word_embeddings=_a , bos_token_id=_a , eos_token_id=_a , **_a )
361
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _A = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys _A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
261
0
'''simple docstring''' # Imports import numpy as np class __magic_name__ : def __init__( self : Union[str, Any] , lowercase_ : Dict=None , lowercase_ : Dict=None , lowercase_ : Optional[Any]=None , lowercase_ : Union[str, Any]=None , lowercase_ : Union[str, Any]=None ): self.set_matricies(red=lowerCamelCase_ , green=lowerCamelCase_ , blue=lowerCamelCase_ , red_edge=lowerCamelCase_ , nir=lowerCamelCase_ ) def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Tuple=None , lowercase_ : List[str]=None , lowercase_ : List[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Union[str, Any]=None ): if red is not None: lowercase_ : int = red if green is not None: lowercase_ : str = green if blue is not None: lowercase_ : Optional[int] = blue if red_edge is not None: lowercase_ : List[Any] = red_edge if nir is not None: lowercase_ : List[Any] = nir return True def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Optional[Any]="" , lowercase_ : Optional[int]=None , lowercase_ : Any=None , lowercase_ : int=None , lowercase_ : Optional[Any]=None , lowercase_ : Tuple=None ): self.set_matricies(red=lowerCamelCase_ , green=lowerCamelCase_ , blue=lowerCamelCase_ , red_edge=lowerCamelCase_ , nir=lowerCamelCase_ ) lowercase_ : Union[str, Any] = { """ARVI2""": self.arvaa, """CCCI""": self.ccci, """CVI""": self.cvi, """GLI""": self.gli, """NDVI""": self.ndvi, """BNDVI""": self.bndvi, """redEdgeNDVI""": self.red_edge_ndvi, """GNDVI""": self.gndvi, """GBNDVI""": self.gbndvi, """GRNDVI""": self.grndvi, """RBNDVI""": self.rbndvi, """PNDVI""": self.pndvi, """ATSAVI""": self.atsavi, """BWDRVI""": self.bwdrvi, """CIgreen""": self.ci_green, """CIrededge""": self.ci_rededge, """CI""": self.ci, """CTVI""": self.ctvi, """GDVI""": self.gdvi, """EVI""": self.evi, """GEMI""": self.gemi, """GOSAVI""": self.gosavi, """GSAVI""": self.gsavi, """Hue""": self.hue, """IVI""": self.ivi, """IPVI""": self.ipvi, """I""": self.i, """RVI""": self.rvi, """MRVI""": self.mrvi, """MSAVI""": self.m_savi, """NormG""": self.norm_g, """NormNIR""": self.norm_nir, """NormR""": self.norm_r, """NGRDI""": self.ngrdi, """RI""": self.ri, """S""": self.s, """IF""": self._if, """DVI""": self.dvi, """TVI""": self.tvi, """NDRE""": self.ndre, } try: return funcs[index]() except KeyError: print("""Index not in the list!""" ) return False def SCREAMING_SNAKE_CASE_ ( self : int ): return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def SCREAMING_SNAKE_CASE_ ( self : Any ): return self.nir * (self.red / (self.green**2)) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def SCREAMING_SNAKE_CASE_ ( self : str ): return (self.nir - self.red) / (self.nir + self.red) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): return (self.nir - self.blue) / (self.nir + self.blue) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): return (self.redEdge - self.red) / (self.redEdge + self.red) def SCREAMING_SNAKE_CASE_ ( self : int ): return (self.nir - self.green) / (self.nir + self.green) def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def SCREAMING_SNAKE_CASE_ ( self : str ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def SCREAMING_SNAKE_CASE_ ( self : Dict ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def SCREAMING_SNAKE_CASE_ ( self : List[str] , lowercase_ : Dict=0.08 , lowercase_ : Optional[int]=1.22 , lowercase_ : List[Any]=0.03 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): return (self.nir / self.green) - 1 def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): return (self.nir / self.redEdge) - 1 def SCREAMING_SNAKE_CASE_ ( self : Dict ): return (self.red - self.blue) / self.red def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : str = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def SCREAMING_SNAKE_CASE_ ( self : int ): return self.nir - self.green def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def SCREAMING_SNAKE_CASE_ ( self : str ): lowercase_ : Tuple = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , lowercase_ : Tuple=0.16 ): return (self.nir - self.green) / (self.nir + self.green + y) def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Union[str, Any]=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def SCREAMING_SNAKE_CASE_ ( self : Any , lowercase_ : List[Any]=None , lowercase_ : Any=None ): return (self.nir - b) / (a * self.red) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return (self.red + self.green + self.blue) / 30.5 def SCREAMING_SNAKE_CASE_ ( self : Dict ): return self.nir / self.red def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return (self.rvi() - 1) / (self.rvi() + 1) def SCREAMING_SNAKE_CASE_ ( self : Any ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return self.green / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): return self.nir / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): return self.red / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): return (self.green - self.red) / (self.green + self.red) def SCREAMING_SNAKE_CASE_ ( self : str ): return (self.red - self.green) / (self.red + self.green) def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : List[str] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) lowercase_ : Union[str, Any] = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): return self.nir / self.red def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): return (self.ndvi() + 0.5) ** (1 / 2) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
239
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : '''simple docstring''' def __init__( self: Any ,lowerCamelCase_: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Any ,lowerCamelCase_: Tuple=0.2 ,lowerCamelCase_: Union[str, Any]=0.2 ) -> List[str]: UpperCAmelCase_ : List[Any] = bp_numa UpperCAmelCase_ : str = bp_numa UpperCAmelCase_ : List[Any] = bp_numa UpperCAmelCase_ : Optional[int] = conva_get[:2] UpperCAmelCase_ : List[Any] = conva_get[2] UpperCAmelCase_ : str = size_pa UpperCAmelCase_ : Optional[int] = rate_w UpperCAmelCase_ : Dict = rate_t UpperCAmelCase_ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] ,self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCAmelCase_ : int = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) UpperCAmelCase_ : int = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) UpperCAmelCase_ : Dict = -2 * np.random.rand(self.conva[1] ) + 1 UpperCAmelCase_ : str = -2 * np.random.rand(self.num_bpa ) + 1 UpperCAmelCase_ : Union[str, Any] = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self: str ,lowerCamelCase_: Optional[Any] ) -> Tuple: # save model dict with pickle UpperCAmelCase_ : Dict = { """num_bp1""": self.num_bpa, """num_bp2""": self.num_bpa, """num_bp3""": self.num_bpa, """conv1""": self.conva, """step_conv1""": self.step_conva, """size_pooling1""": self.size_poolinga, """rate_weight""": self.rate_weight, """rate_thre""": self.rate_thre, """w_conv1""": self.w_conva, """wkj""": self.wkj, """vji""": self.vji, """thre_conv1""": self.thre_conva, """thre_bp2""": self.thre_bpa, """thre_bp3""": self.thre_bpa, } with open(lowerCamelCase_ ,"""wb""" ) as f: pickle.dump(lowerCamelCase_ ,lowerCamelCase_ ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls: List[str] ,lowerCamelCase_: str ) -> List[str]: # read saved model with open(lowerCamelCase_ ,"""rb""" ) as f: UpperCAmelCase_ : Any = pickle.load(lowerCamelCase_ ) # noqa: S301 UpperCAmelCase_ : Union[str, Any] = model_dic.get("""conv1""" ) conv_get.append(model_dic.get("""step_conv1""" ) ) UpperCAmelCase_ : List[str] = model_dic.get("""size_pooling1""" ) UpperCAmelCase_ : Tuple = model_dic.get("""num_bp1""" ) UpperCAmelCase_ : Optional[Any] = model_dic.get("""num_bp2""" ) UpperCAmelCase_ : List[str] = model_dic.get("""num_bp3""" ) UpperCAmelCase_ : List[Any] = model_dic.get("""rate_weight""" ) UpperCAmelCase_ : Dict = model_dic.get("""rate_thre""" ) # create model instance UpperCAmelCase_ : List[Any] = CNN(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) # modify model parameter UpperCAmelCase_ : Any = model_dic.get("""w_conv1""" ) UpperCAmelCase_ : int = model_dic.get("""wkj""" ) UpperCAmelCase_ : int = model_dic.get("""vji""" ) UpperCAmelCase_ : Optional[int] = model_dic.get("""thre_conv1""" ) UpperCAmelCase_ : List[str] = model_dic.get("""thre_bp2""" ) UpperCAmelCase_ : Dict = model_dic.get("""thre_bp3""" ) return conv_ins def A__ ( self: List[Any] ,lowerCamelCase_: Union[str, Any] ) -> Tuple: return 1 / (1 + np.exp(-1 * x )) def A__ ( self: Union[str, Any] ,lowerCamelCase_: Union[str, Any] ) -> Optional[Any]: return round(lowerCamelCase_ ,3 ) def A__ ( self: Tuple ,lowerCamelCase_: Any ,lowerCamelCase_: List[str] ,lowerCamelCase_: str ,lowerCamelCase_: Any ,lowerCamelCase_: Union[str, Any] ) -> Any: # convolution process UpperCAmelCase_ : Optional[Any] = convs[0] UpperCAmelCase_ : int = convs[1] UpperCAmelCase_ : int = np.shape(lowerCamelCase_ )[0] # get the data slice of original image data, data_focus UpperCAmelCase_ : Dict = [] for i_focus in range(0 ,size_data - size_conv + 1 ,lowerCamelCase_ ): for j_focus in range(0 ,size_data - size_conv + 1 ,lowerCamelCase_ ): UpperCAmelCase_ : Union[str, Any] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(lowerCamelCase_ ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCAmelCase_ : Any = [] UpperCAmelCase_ : Tuple = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(lowerCamelCase_ ): UpperCAmelCase_ : Optional[int] = [] for i_focus in range(len(lowerCamelCase_ ) ): UpperCAmelCase_ : int = ( np.sum(np.multiply(data_focus[i_focus] ,w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(lowerCamelCase_ ) ) UpperCAmelCase_ : Union[str, Any] = np.asmatrix(lowerCamelCase_ ).reshape( lowerCamelCase_ ,lowerCamelCase_ ) data_featuremap.append(lowerCamelCase_ ) # expanding the data slice to One dimenssion UpperCAmelCase_ : Optional[Any] = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(lowerCamelCase_ ) ) UpperCAmelCase_ : Optional[int] = np.asarray(lowerCamelCase_ ) return focus_list, data_featuremap def A__ ( self: Tuple ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Optional[Any]="average_pool" ) -> List[Any]: # pooling process UpperCAmelCase_ : Optional[Any] = len(featuremaps[0] ) UpperCAmelCase_ : Any = int(size_map / size_pooling ) UpperCAmelCase_ : Optional[int] = [] for i_map in range(len(lowerCamelCase_ ) ): UpperCAmelCase_ : Any = featuremaps[i_map] UpperCAmelCase_ : Tuple = [] for i_focus in range(0 ,lowerCamelCase_ ,lowerCamelCase_ ): for j_focus in range(0 ,lowerCamelCase_ ,lowerCamelCase_ ): UpperCAmelCase_ : str = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(lowerCamelCase_ ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(lowerCamelCase_ ) ) UpperCAmelCase_ : int = np.asmatrix(lowerCamelCase_ ).reshape(lowerCamelCase_ ,lowerCamelCase_ ) featuremap_pooled.append(lowerCamelCase_ ) return featuremap_pooled def A__ ( self: Union[str, Any] ,lowerCamelCase_: Tuple ) -> Optional[int]: # expanding three dimension data to one dimension list UpperCAmelCase_ : List[Any] = [] for i in range(len(lowerCamelCase_ ) ): UpperCAmelCase_ : Tuple = np.shape(data[i] ) UpperCAmelCase_ : Optional[int] = data[i].reshape(1 ,shapes[0] * shapes[1] ) UpperCAmelCase_ : Optional[int] = data_listed.getA().tolist()[0] data_expanded.extend(lowerCamelCase_ ) UpperCAmelCase_ : int = np.asarray(lowerCamelCase_ ) return data_expanded def A__ ( self: Optional[Any] ,lowerCamelCase_: Optional[int] ) -> Union[str, Any]: # expanding matrix to one dimension list UpperCAmelCase_ : List[Any] = np.asarray(lowerCamelCase_ ) UpperCAmelCase_ : str = np.shape(lowerCamelCase_ ) UpperCAmelCase_ : Dict = data_mat.reshape(1 ,shapes[0] * shapes[1] ) return data_expanded def A__ ( self: str ,lowerCamelCase_: Dict ,lowerCamelCase_: int ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Any ) -> Union[str, Any]: UpperCAmelCase_ : Any = [] UpperCAmelCase_ : Tuple = 0 for i_map in range(lowerCamelCase_ ): UpperCAmelCase_ : Optional[Any] = np.ones((size_map, size_map) ) for i in range(0 ,lowerCamelCase_ ,lowerCamelCase_ ): for j in range(0 ,lowerCamelCase_ ,lowerCamelCase_ ): UpperCAmelCase_ : Any = pd_pool[ i_pool ] UpperCAmelCase_ : List[str] = i_pool + 1 UpperCAmelCase_ : Optional[Any] = np.multiply( lowerCamelCase_ ,np.multiply(out_map[i_map] ,(1 - out_map[i_map]) ) ) pd_all.append(lowerCamelCase_ ) return pd_all def A__ ( self: str ,lowerCamelCase_: int ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Any ,lowerCamelCase_: List[str] ,lowerCamelCase_: Any=bool ) -> Optional[int]: # model traning print("""----------------------Start Training-------------------------""" ) print((""" - - Shape: Train_Data """, np.shape(lowerCamelCase_ )) ) print((""" - - Shape: Teach_Data """, np.shape(lowerCamelCase_ )) ) UpperCAmelCase_ : str = 0 UpperCAmelCase_ : Tuple = [] UpperCAmelCase_ : Any = 10000 while rp < n_repeat and mse >= error_accuracy: UpperCAmelCase_ : List[str] = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(lowerCamelCase_ ) ): # print('------------Learning Image: %d--------------'%p) UpperCAmelCase_ : str = np.asmatrix(datas_train[p] ) UpperCAmelCase_ : Optional[Any] = np.asarray(datas_teach[p] ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.convolute( lowerCamelCase_ ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase_ : List[Any] = self.pooling(lowerCamelCase_ ,self.size_poolinga ) UpperCAmelCase_ : int = np.shape(lowerCamelCase_ ) UpperCAmelCase_ : Dict = self._expand(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = data_bp_input UpperCAmelCase_ : Optional[Any] = np.dot(lowerCamelCase_ ,self.vji.T ) - self.thre_bpa UpperCAmelCase_ : int = self.sig(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = np.dot(lowerCamelCase_ ,self.wkj.T ) - self.thre_bpa UpperCAmelCase_ : Optional[Any] = self.sig(lowerCamelCase_ ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCAmelCase_ : List[str] = np.multiply( (data_teach - bp_outa) ,np.multiply(lowerCamelCase_ ,(1 - bp_outa) ) ) UpperCAmelCase_ : List[Any] = np.multiply( np.dot(lowerCamelCase_ ,self.wkj ) ,np.multiply(lowerCamelCase_ ,(1 - bp_outa) ) ) UpperCAmelCase_ : Any = np.dot(lowerCamelCase_ ,self.vji ) UpperCAmelCase_ : Tuple = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCAmelCase_ : List[str] = pd_conva_pooled.T.getA().tolist() UpperCAmelCase_ : str = self._calculate_gradient_from_pool( lowerCamelCase_ ,lowerCamelCase_ ,shape_featuremapa[0] ,shape_featuremapa[1] ,self.size_poolinga ,) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCAmelCase_ : List[str] = self._expand_mat(pd_conva_all[k_conv] ) UpperCAmelCase_ : Optional[Any] = self.rate_weight * np.dot(lowerCamelCase_ ,lowerCamelCase_ ) UpperCAmelCase_ : int = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCAmelCase_ : str = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCAmelCase_ : int = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCAmelCase_ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCAmelCase_ : int = self.thre_bpa - pd_k_all * self.rate_thre UpperCAmelCase_ : str = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCAmelCase_ : int = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCAmelCase_ : int = rp + 1 UpperCAmelCase_ : Any = error_count / patterns all_mse.append(lowerCamelCase_ ) def draw_error(): UpperCAmelCase_ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(lowerCamelCase_ ,"""+-""" ) plt.plot(lowerCamelCase_ ,"""r--""" ) plt.xlabel("""Learning Times""" ) plt.ylabel("""All_mse""" ) plt.grid(lowerCamelCase_ ,alpha=0.5 ) plt.show() print("""------------------Training Complished---------------------""" ) print((""" - - Training epoch: """, rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self: Optional[int] ,lowerCamelCase_: Any ) -> Tuple: # model predict UpperCAmelCase_ : Union[str, Any] = [] print("""-------------------Start Testing-------------------------""" ) print((""" - - Shape: Test_Data """, np.shape(lowerCamelCase_ )) ) for p in range(len(lowerCamelCase_ ) ): UpperCAmelCase_ : int = np.asmatrix(datas_test[p] ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.convolute( lowerCamelCase_ ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase_ : Optional[Any] = self.pooling(lowerCamelCase_ ,self.size_poolinga ) UpperCAmelCase_ : str = self._expand(lowerCamelCase_ ) UpperCAmelCase_ : str = data_bp_input UpperCAmelCase_ : Union[str, Any] = bp_outa * self.vji.T - self.thre_bpa UpperCAmelCase_ : Optional[int] = self.sig(lowerCamelCase_ ) UpperCAmelCase_ : Tuple = bp_outa * self.wkj.T - self.thre_bpa UpperCAmelCase_ : List[Any] = self.sig(lowerCamelCase_ ) produce_out.extend(bp_outa.getA().tolist() ) UpperCAmelCase_ : int = [list(map(self.do_round ,lowerCamelCase_ ) ) for each in produce_out] return np.asarray(lowerCamelCase_ ) def A__ ( self: Optional[Any] ,lowerCamelCase_: Dict ) -> Tuple: # return the data of image after convoluting process so we can check it out UpperCAmelCase_ : Optional[int] = np.asmatrix(lowerCamelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.convolute( lowerCamelCase_ ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase_ : Dict = self.pooling(lowerCamelCase_ ,self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
345
0
'''simple docstring''' import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py lowerCAmelCase :List[str] = '''.''' if __name__ == "__main__": lowerCAmelCase :Optional[int] = os.path.join(REPO_PATH, '''utils/documentation_tests.txt''') lowerCAmelCase :List[Any] = [] lowerCAmelCase :Dict = [] with open(doctest_file_path) as fp: for line in fp: lowerCAmelCase :Dict = line.strip() lowerCAmelCase :Tuple = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: lowerCAmelCase :List[str] = '''\n'''.join(non_existent_paths) raise ValueError(F'`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}') if all_paths != sorted(all_paths): raise ValueError('''Files in `utils/documentation_tests.txt` are not in alphabetical order.''')
275
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. lowerCAmelCase :Optional[int] = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCamelCase ( lowerCAmelCase : Tuple ): """simple docstring""" config.addinivalue_line( 'markers' , 'is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested' ) config.addinivalue_line( 'markers' , 'is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested' ) config.addinivalue_line('markers' , 'is_pipeline_test: mark test to run only when pipelines are tested' ) config.addinivalue_line('markers' , 'is_staging_test: mark test to run only in the staging environment' ) config.addinivalue_line('markers' , 'accelerate_tests: mark test that require accelerate' ) config.addinivalue_line('markers' , 'tool_tests: mark the tool tests that are run on their specific schedule' ) def lowerCamelCase ( lowerCAmelCase : Any ): """simple docstring""" from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowerCAmelCase ) def lowerCamelCase ( lowerCAmelCase : Dict ): """simple docstring""" from transformers.testing_utils import pytest_terminal_summary_main __magic_name__ : Tuple = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(lowerCAmelCase , id=lowerCAmelCase ) def lowerCamelCase ( lowerCAmelCase : List[Any] , lowerCAmelCase : Tuple ): """simple docstring""" if exitstatus == 5: __magic_name__ : Any = 0 # Doctest custom flag to ignore output. lowerCAmelCase :List[str] = doctest.register_optionflag('''IGNORE_RESULT''') lowerCAmelCase :Union[str, Any] = doctest.OutputChecker class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __lowerCAmelCase ( self : List[str] , _A : Tuple , _A : Tuple , _A : str ) -> int: if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , _A , _A , _A ) lowerCAmelCase :Optional[Any] = CustomOutputChecker lowerCAmelCase :int = HfDoctestModule lowerCAmelCase :Any = HfDocTestParser
275
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "gpt_neox" def __init__(self , UpperCAmelCase=50432 , UpperCAmelCase=6144 , UpperCAmelCase=44 , UpperCAmelCase=64 , UpperCAmelCase=24576 , UpperCAmelCase="gelu" , UpperCAmelCase=0.25 , UpperCAmelCase=10000 , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase=2048 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=0 , UpperCAmelCase=2 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=None , **UpperCAmelCase , ) -> Optional[Any]: super().__init__(bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) _snake_case = vocab_size _snake_case = max_position_embeddings _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = rotary_pct _snake_case = rotary_emb_base _snake_case = attention_dropout _snake_case = hidden_dropout _snake_case = classifier_dropout _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = use_parallel_residual _snake_case = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( """The hidden size is not divisble by the number of attention heads! Make sure to update them!""" ) def lowercase (self ) -> Union[str, Any]: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , UpperCAmelCase ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f"""got {self.rope_scaling}""" ) _snake_case = self.rope_scaling.get("""type""" , UpperCAmelCase ) _snake_case = self.rope_scaling.get("""factor""" , UpperCAmelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(UpperCAmelCase , UpperCAmelCase ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
341
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
1
"""simple docstring""" from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline _UpperCamelCase = logging.get_logger(__name__) class lowerCamelCase__ ( __UpperCamelCase ): def _UpperCamelCase ( self ,A ): if isinstance(A ,A ): UpperCAmelCase = [label.strip() for label in labels.split(""",""" ) if label.strip()] return labels def __call__( self ,A ,A ,A ): if len(A ) == 0 or len(A ) == 0: raise ValueError("""You must include at least one label and at least one sequence.""" ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( """The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. """ """Make sure the passed template includes formatting syntax such as {{}} where the label should go.""" ).format(A ) ) if isinstance(A ,A ): UpperCAmelCase = [sequences] UpperCAmelCase = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(A )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(__UpperCamelCase ) class lowerCamelCase__ ( __UpperCamelCase ): def __init__( self ,A=ZeroShotClassificationArgumentHandler() ,*A ,**A ): UpperCAmelCase = args_parser super().__init__(*A ,**A ) if self.entailment_id == -1: logger.warning( """Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to """ """-1. Define a descriptive label2id mapping in the model config to ensure correct outputs.""" ) @property def _UpperCamelCase ( self ): for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("""entail""" ): return ind return -1 def _UpperCamelCase ( self ,A ,A=True ,A=True ,A=TruncationStrategy.ONLY_FIRST ,**A ): UpperCAmelCase = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( """Tokenizer was not supporting padding necessary for zero-shot, attempting to use """ """ `pad_token=eos_token`""" ) UpperCAmelCase = self.tokenizer.eos_token try: UpperCAmelCase = self.tokenizer( A ,add_special_tokens=A ,return_tensors=A ,padding=A ,truncation=A ,) except Exception as e: if "too short" in str(A ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. UpperCAmelCase = self.tokenizer( A ,add_special_tokens=A ,return_tensors=A ,padding=A ,truncation=TruncationStrategy.DO_NOT_TRUNCATE ,) else: raise e return inputs def _UpperCamelCase ( self ,**A ): if kwargs.get("""multi_class""" ,A ) is not None: UpperCAmelCase = kwargs["""multi_class"""] logger.warning( """The `multi_class` argument has been deprecated and renamed to `multi_label`. """ """`multi_class` will be removed in a future version of Transformers.""" ) UpperCAmelCase = {} if "candidate_labels" in kwargs: UpperCAmelCase = self._args_parser._parse_labels(kwargs["""candidate_labels"""] ) if "hypothesis_template" in kwargs: UpperCAmelCase = kwargs["""hypothesis_template"""] UpperCAmelCase = {} if "multi_label" in kwargs: UpperCAmelCase = kwargs["""multi_label"""] return preprocess_params, {}, postprocess_params def __call__( self ,A ,*A ,**A ,): if len(A ) == 0: pass elif len(A ) == 1 and "candidate_labels" not in kwargs: UpperCAmelCase = args[0] else: raise ValueError(F'''Unable to understand extra arguments {args}''' ) return super().__call__(A ,**A ) def _UpperCamelCase ( self ,A ,A=None ,A="This example is {}." ): UpperCAmelCase = self._args_parser(A ,A ,A ) for i, (candidate_label, sequence_pair) in enumerate(zip(A ,A ) ): UpperCAmelCase = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(A ) - 1, **model_input, } def _UpperCamelCase ( self ,A ): UpperCAmelCase = inputs["""candidate_label"""] UpperCAmelCase = inputs["""sequence"""] UpperCAmelCase = {k: inputs[k] for k in self.tokenizer.model_input_names} UpperCAmelCase = self.model(**A ) UpperCAmelCase = { """candidate_label""": candidate_label, """sequence""": sequence, """is_last""": inputs["""is_last"""], **outputs, } return model_outputs def _UpperCamelCase ( self ,A ,A=False ): UpperCAmelCase = [outputs["""candidate_label"""] for outputs in model_outputs] UpperCAmelCase = [outputs["""sequence"""] for outputs in model_outputs] UpperCAmelCase = np.concatenate([output["""logits"""].numpy() for output in model_outputs] ) UpperCAmelCase = logits.shape[0] UpperCAmelCase = len(A ) UpperCAmelCase = N // n UpperCAmelCase = logits.reshape((num_sequences, n, -1) ) if multi_label or len(A ) == 1: # softmax over the entailment vs. contradiction dim for each label independently UpperCAmelCase = self.entailment_id UpperCAmelCase = -1 if entailment_id == 0 else 0 UpperCAmelCase = reshaped_outputs[..., [contradiction_id, entailment_id]] UpperCAmelCase = np.exp(A ) / np.exp(A ).sum(-1 ,keepdims=A ) UpperCAmelCase = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels UpperCAmelCase = reshaped_outputs[..., self.entailment_id] UpperCAmelCase = np.exp(A ) / np.exp(A ).sum(-1 ,keepdims=A ) UpperCAmelCase = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
356
"""simple docstring""" import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( """The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion""" ) _UpperCamelCase = None _UpperCamelCase = { """7B""": 11008, """13B""": 13824, """30B""": 17920, """65B""": 22016, """70B""": 28672, } _UpperCamelCase = { """7B""": 1, """7Bf""": 1, """13B""": 2, """13Bf""": 2, """30B""": 4, """65B""": 8, """70B""": 8, """70Bf""": 8, } def _a ( _snake_case , _snake_case=1 , _snake_case=256 ): """simple docstring""" return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def _a ( _snake_case ): """simple docstring""" with open(_snake_case , """r""" ) as f: return json.load(_snake_case ) def _a ( _snake_case , _snake_case ): """simple docstring""" with open(_snake_case , """w""" ) as f: json.dump(_snake_case , _snake_case ) def _a ( _snake_case , _snake_case , _snake_case , _snake_case=True ): """simple docstring""" os.makedirs(_snake_case , exist_ok=_snake_case ) UpperCAmelCase = os.path.join(_snake_case , """tmp""" ) os.makedirs(_snake_case , exist_ok=_snake_case ) UpperCAmelCase = read_json(os.path.join(_snake_case , """params.json""" ) ) UpperCAmelCase = NUM_SHARDS[model_size] UpperCAmelCase = params["""n_layers"""] UpperCAmelCase = params["""n_heads"""] UpperCAmelCase = n_heads // num_shards UpperCAmelCase = params["""dim"""] UpperCAmelCase = dim // n_heads UpperCAmelCase = 10000.0 UpperCAmelCase = 1.0 / (base ** (torch.arange(0 , _snake_case , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase = params["""n_kv_heads"""] # for GQA / MQA UpperCAmelCase = n_heads_per_shard // num_key_value_heads UpperCAmelCase = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase = n_heads UpperCAmelCase = n_heads_per_shard UpperCAmelCase = dim # permute for sliced rotary def permute(_snake_case , _snake_case=n_heads , _snake_case=dim , _snake_case=dim ): return w.view(_snake_case , dima // n_heads // 2 , 2 , _snake_case ).transpose(1 , 2 ).reshape(_snake_case , _snake_case ) print(F'''Fetching all parameters from the checkpoint at {input_base_path}.''' ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase = torch.load(os.path.join(_snake_case , """consolidated.00.pth""" ) , map_location="""cpu""" ) else: # Sharded UpperCAmelCase = [ torch.load(os.path.join(_snake_case , F'''consolidated.{i:02d}.pth''' ) , map_location="""cpu""" ) for i in range(_snake_case ) ] UpperCAmelCase = 0 UpperCAmelCase = {"""weight_map""": {}} for layer_i in range(_snake_case ): UpperCAmelCase = F'''pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin''' if model_size == "7B": # Unsharded UpperCAmelCase = { F'''model.layers.{layer_i}.self_attn.q_proj.weight''': permute( loaded[F'''layers.{layer_i}.attention.wq.weight'''] ), F'''model.layers.{layer_i}.self_attn.k_proj.weight''': permute( loaded[F'''layers.{layer_i}.attention.wk.weight'''] ), F'''model.layers.{layer_i}.self_attn.v_proj.weight''': loaded[F'''layers.{layer_i}.attention.wv.weight'''], F'''model.layers.{layer_i}.self_attn.o_proj.weight''': loaded[F'''layers.{layer_i}.attention.wo.weight'''], F'''model.layers.{layer_i}.mlp.gate_proj.weight''': loaded[F'''layers.{layer_i}.feed_forward.w1.weight'''], F'''model.layers.{layer_i}.mlp.down_proj.weight''': loaded[F'''layers.{layer_i}.feed_forward.w2.weight'''], F'''model.layers.{layer_i}.mlp.up_proj.weight''': loaded[F'''layers.{layer_i}.feed_forward.w3.weight'''], F'''model.layers.{layer_i}.input_layernorm.weight''': loaded[F'''layers.{layer_i}.attention_norm.weight'''], F'''model.layers.{layer_i}.post_attention_layernorm.weight''': loaded[F'''layers.{layer_i}.ffn_norm.weight'''], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase = { F'''model.layers.{layer_i}.input_layernorm.weight''': loaded[0][ F'''layers.{layer_i}.attention_norm.weight''' ].clone(), F'''model.layers.{layer_i}.post_attention_layernorm.weight''': loaded[0][ F'''layers.{layer_i}.ffn_norm.weight''' ].clone(), } UpperCAmelCase = permute( torch.cat( [ loaded[i][F'''layers.{layer_i}.attention.wq.weight'''].view(_snake_case , _snake_case , _snake_case ) for i in range(_snake_case ) ] , dim=0 , ).reshape(_snake_case , _snake_case ) ) UpperCAmelCase = permute( torch.cat( [ loaded[i][F'''layers.{layer_i}.attention.wk.weight'''].view( _snake_case , _snake_case , _snake_case ) for i in range(_snake_case ) ] , dim=0 , ).reshape(_snake_case , _snake_case ) , _snake_case , _snake_case , _snake_case , ) UpperCAmelCase = torch.cat( [ loaded[i][F'''layers.{layer_i}.attention.wv.weight'''].view( _snake_case , _snake_case , _snake_case ) for i in range(_snake_case ) ] , dim=0 , ).reshape(_snake_case , _snake_case ) UpperCAmelCase = torch.cat( [loaded[i][F'''layers.{layer_i}.attention.wo.weight'''] for i in range(_snake_case )] , dim=1 ) UpperCAmelCase = torch.cat( [loaded[i][F'''layers.{layer_i}.feed_forward.w1.weight'''] for i in range(_snake_case )] , dim=0 ) UpperCAmelCase = torch.cat( [loaded[i][F'''layers.{layer_i}.feed_forward.w2.weight'''] for i in range(_snake_case )] , dim=1 ) UpperCAmelCase = torch.cat( [loaded[i][F'''layers.{layer_i}.feed_forward.w3.weight'''] for i in range(_snake_case )] , dim=0 ) UpperCAmelCase = inv_freq for k, v in state_dict.items(): UpperCAmelCase = filename param_count += v.numel() torch.save(_snake_case , os.path.join(_snake_case , _snake_case ) ) UpperCAmelCase = F'''pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin''' if model_size == "7B": # Unsharded UpperCAmelCase = { """model.embed_tokens.weight""": loaded["""tok_embeddings.weight"""], """model.norm.weight""": loaded["""norm.weight"""], """lm_head.weight""": loaded["""output.weight"""], } else: UpperCAmelCase = { """model.norm.weight""": loaded[0]["""norm.weight"""], """model.embed_tokens.weight""": torch.cat( [loaded[i]["""tok_embeddings.weight"""] for i in range(_snake_case )] , dim=1 ), """lm_head.weight""": torch.cat([loaded[i]["""output.weight"""] for i in range(_snake_case )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase = filename param_count += v.numel() torch.save(_snake_case , os.path.join(_snake_case , _snake_case ) ) # Write configs UpperCAmelCase = {"""total_size""": param_count * 2} write_json(_snake_case , os.path.join(_snake_case , """pytorch_model.bin.index.json""" ) ) UpperCAmelCase = params["""ffn_dim_multiplier"""] if """ffn_dim_multiplier""" in params else 1 UpperCAmelCase = params["""multiple_of"""] if """multiple_of""" in params else 256 UpperCAmelCase = LlamaConfig( hidden_size=_snake_case , intermediate_size=compute_intermediate_size(_snake_case , _snake_case , _snake_case ) , num_attention_heads=params["""n_heads"""] , num_hidden_layers=params["""n_layers"""] , rms_norm_eps=params["""norm_eps"""] , num_key_value_heads=_snake_case , ) config.save_pretrained(_snake_case ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print("""Loading the checkpoint in a Llama model.""" ) UpperCAmelCase = LlamaForCausalLM.from_pretrained(_snake_case , torch_dtype=torch.floataa , low_cpu_mem_usage=_snake_case ) # Avoid saving this as part of the config. del model.config._name_or_path print("""Saving in the Transformers format.""" ) model.save_pretrained(_snake_case , safe_serialization=_snake_case ) shutil.rmtree(_snake_case ) def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(F'''Saving a {tokenizer_class.__name__} to {tokenizer_path}.''' ) UpperCAmelCase = tokenizer_class(_snake_case ) tokenizer.save_pretrained(_snake_case ) def _a ( ): """simple docstring""" UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """--input_dir""" , help="""Location of LLaMA weights, which contains tokenizer.model and model folders""" , ) parser.add_argument( """--model_size""" , choices=["""7B""", """7Bf""", """13B""", """13Bf""", """30B""", """65B""", """70B""", """70Bf""", """tokenizer_only"""] , ) parser.add_argument( """--output_dir""" , help="""Location to write HF model and tokenizer""" , ) parser.add_argument("""--safe_serialization""" , type=_snake_case , help="""Whether or not to save using `safetensors`.""" ) UpperCAmelCase = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase = os.path.join(args.input_dir , """tokenizer.model""" ) write_tokenizer(args.output_dir , _snake_case ) if __name__ == "__main__": main()
234
0
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __A (_SCREAMING_SNAKE_CASE ) ->str: """simple docstring""" lowerCAmelCase__ :int = checkpoints.load_tax_checkpoint(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[Any] = flatten_dict(_SCREAMING_SNAKE_CASE ) return flax_params def __A (_SCREAMING_SNAKE_CASE ) ->Any: """simple docstring""" lowerCAmelCase__ :Any = {} lowerCAmelCase__ :List[str] = { 'token_embedder': 'embeddings', 'encoder_norm': 'layernorm', 'kernel': 'weight', '.out': '.output', 'scale': 'weight', 'embedders_0.pos_embedding': 'row_embedder.weight', 'embedders_1.pos_embedding': 'column_embedder.weight', } lowerCAmelCase__ :str = { 'query': 'attention.query', 'key': 'attention.key', 'value': 'attention.value', 'output.dense': 'output', 'encoder_decoder_attention.o': 'encoder_decoder_attention.attention.o', 'pre_self_attention_layer_norm': 'self_attention.layer_norm', 'pre_cross_attention_layer_norm': 'encoder_decoder_attention.layer_norm', 'mlp.': 'mlp.DenseReluDense.', 'pre_mlp_layer_norm': 'mlp.layer_norm', 'self_attention.o': 'self_attention.attention.o', 'decoder.embeddings.embedding': 'decoder.embed_tokens.weight', 'decoder.relpos_bias.rel_embedding': 'decoder.layer.0.self_attention.attention.relative_attention_bias.weight', 'decoder.decoder_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.logits_dense.weight': 'decoder.lm_head.weight', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowerCAmelCase__ :List[Any] = '.'.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowerCAmelCase__ :Any = new_key.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowerCAmelCase__ :List[Any] = new_key.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowerCAmelCase__ :int = re.sub(r'layers_(\d+)' , r'layer.\1' , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = new_key.replace('encoder' , 'encoder.encoder' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowerCAmelCase__ :Dict = re.sub(r'layers_(\d+)' , r'layer.\1' , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = flax_dict[key] lowerCAmelCase__ :List[str] = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowerCAmelCase__ :Any = torch.from_numpy(converted_dict[key].T ) else: lowerCAmelCase__ :List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False ) ->str: """simple docstring""" lowerCAmelCase__ :Any = get_flax_param(_SCREAMING_SNAKE_CASE ) if not use_large: lowerCAmelCase__ :List[str] = PixaStructVisionConfig() lowerCAmelCase__ :Tuple = PixaStructTextConfig() else: lowerCAmelCase__ :Union[str, Any] = PixaStructVisionConfig( hidden_size=1536 , d_ff=3968 , num_attention_heads=24 , num_hidden_layers=18 ) lowerCAmelCase__ :List[str] = PixaStructTextConfig(hidden_size=1536 , d_ff=3968 , num_heads=24 , num_layers=18 ) lowerCAmelCase__ :Union[str, Any] = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = PixaStructForConditionalGeneration(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Union[str, Any] = rename_and_convert_flax_params(_SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Tuple = AutoTokenizer.from_pretrained('ybelkada/test-pix2struct-tokenizer' ) lowerCAmelCase__ :Union[str, Any] = PixaStructImageProcessor() lowerCAmelCase__ :Union[str, Any] = PixaStructProcessor(image_processor=_SCREAMING_SNAKE_CASE , tokenizer=_SCREAMING_SNAKE_CASE ) if use_large: lowerCAmelCase__ :Tuple = 4096 lowerCAmelCase__ :Dict = True # mkdir if needed os.makedirs(_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) processor.save_pretrained(_SCREAMING_SNAKE_CASE ) print('Model saved in {}'.format(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("""--t5x_checkpoint_path""", default=None, type=str, help="""Path to the original T5x checkpoint.""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--use_large""", action="""store_true""", help="""Use large model.""") parser.add_argument("""--is_vqa""", action="""store_true""", help="""Use large model.""") __A = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
293
"""simple docstring""" import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class _lowerCAmelCase : """simple docstring""" @staticmethod def snake_case ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def __A (_SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. __A = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" __magic_name__ :str = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Dict = pipeline( 'document-question-answering' , model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) lowerCAmelCase__ :Dict = INVOICE_URL lowerCAmelCase__ :Dict = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '' ) ) ) lowerCAmelCase__ :List[Any] = 'What is the placebo?' lowerCAmelCase__ :Dict = [ { 'image': load_image(__UpperCAmelCase ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :int = dqa_pipeline(__UpperCAmelCase , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {'score': ANY(__UpperCAmelCase ), 'answer': ANY(__UpperCAmelCase ), 'start': ANY(__UpperCAmelCase ), 'end': ANY(__UpperCAmelCase )}, {'score': ANY(__UpperCAmelCase ), 'answer': ANY(__UpperCAmelCase ), 'start': ANY(__UpperCAmelCase ), 'end': ANY(__UpperCAmelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) lowerCAmelCase__ :Union[str, Any] = INVOICE_URL lowerCAmelCase__ :Tuple = 'How many cats are there?' lowerCAmelCase__ :List[str] = [ {'score': 0.00_01, 'answer': 'oy 2312/2019', 'start': 3_8, 'end': 3_9}, {'score': 0.00_01, 'answer': 'oy 2312/2019 DUE', 'start': 3_8, 'end': 4_0}, ] lowerCAmelCase__ :Any = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) lowerCAmelCase__ :Any = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowerCAmelCase__ :List[Any] = './tests/fixtures/tests_samples/COCO/000000039769.png' lowerCAmelCase__ :List[Any] = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) # We can optionnally pass directly the words and bounding boxes lowerCAmelCase__ :Dict = './tests/fixtures/tests_samples/COCO/000000039769.png' lowerCAmelCase__ :List[str] = [] lowerCAmelCase__ :int = [] lowerCAmelCase__ :List[str] = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , words=__UpperCAmelCase , boxes=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) lowerCAmelCase__ :str = INVOICE_URL lowerCAmelCase__ :List[Any] = 'What is the invoice number?' lowerCAmelCase__ :Tuple = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.00_09, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) lowerCAmelCase__ :Union[str, Any] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.00_09, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) lowerCAmelCase__ :Dict = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.99_44, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.00_09, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=5_0 , ) lowerCAmelCase__ :List[Any] = INVOICE_URL lowerCAmelCase__ :List[Any] = 'What is the invoice number?' lowerCAmelCase__ :Optional[Any] = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, {'score': 0.99_48, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) lowerCAmelCase__ :List[str] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, {'score': 0.99_48, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) lowerCAmelCase__ :int = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.99_74, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, {'score': 0.99_48, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=__UpperCAmelCase , revision='3dc6de3' , ) lowerCAmelCase__ :List[str] = INVOICE_URL lowerCAmelCase__ :Any = 'What is the invoice number?' lowerCAmelCase__ :List[Any] = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.08_19, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, ] , ) lowerCAmelCase__ :Optional[int] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.08_19, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, ] , ) lowerCAmelCase__ :List[str] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.42_51, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.08_19, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, ] ] * 2 , ) lowerCAmelCase__ :Dict = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '' ) ) ) # This model should also work if `image` is set to None lowerCAmelCase__ :Tuple = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.08_19, 'answer': '1110212019', 'start': 2_3, 'end': 2_3}, ] , ) @slow @require_torch @require_pytesseract @require_vision def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ :Tuple = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=__UpperCAmelCase , revision='3dc6de3' , max_seq_len=5_0 , ) lowerCAmelCase__ :Dict = INVOICE_URL lowerCAmelCase__ :List[Any] = 'What is the invoice number?' lowerCAmelCase__ :List[str] = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.99_98, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) lowerCAmelCase__ :List[str] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.99_99, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.99_98, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] ] * 2 , ) lowerCAmelCase__ :Optional[Any] = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '' ) ) ) # This model should also work if `image` is set to None lowerCAmelCase__ :List[str] = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, {'score': 0.99_98, 'answer': 'us-001', 'start': 1_6, 'end': 1_6}, ] , ) @slow @require_torch def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) lowerCAmelCase__ :Dict = INVOICE_URL lowerCAmelCase__ :str = 'What is the invoice number?' lowerCAmelCase__ :Tuple = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def snake_case ( self ): '''simple docstring''' pass
293
1
import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def a ( lowerCamelCase__ ): '''simple docstring''' return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue A_ : Optional[int] = key.replace("""heads.cmd.mim_head.cls.predictions""" , """mmm_image_head""" ) A_ : str = key.replace("""heads.cmd.mlm_head.cls.predictions""" , """mmm_text_head""" ) A_ : List[Any] = key.replace("""heads.cmd.itm_head.cls""" , """itm_head""" ) A_ : List[str] = key.replace("""heads.cmd.itm_head.pooler""" , """itm_head.pooler""" ) A_ : str = key.replace("""heads.cmd.clip_head.logit_scale""" , """flava.logit_scale""" ) A_ : List[str] = key.replace("""heads.fairseq_mlm.cls.predictions""" , """mlm_head""" ) A_ : Tuple = key.replace("""heads.imagenet.mim_head.cls.predictions""" , """mim_head""" ) A_ : Tuple = key.replace("""mm_text_projection""" , """flava.text_to_mm_projection""" ) A_ : Any = key.replace("""mm_image_projection""" , """flava.image_to_mm_projection""" ) A_ : Tuple = key.replace("""image_encoder.module""" , """flava.image_model""" ) A_ : Dict = key.replace("""text_encoder.module""" , """flava.text_model""" ) A_ : int = key.replace("""mm_encoder.module.encoder.cls_token""" , """flava.multimodal_model.cls_token""" ) A_ : Tuple = key.replace("""mm_encoder.module""" , """flava.multimodal_model""" ) A_ : Optional[Any] = key.replace("""text_projection""" , """flava.text_projection""" ) A_ : Any = key.replace("""image_projection""" , """flava.image_projection""" ) A_ : int = value.float() for key, value in codebook_state_dict.items(): A_ : Dict = value return upgrade @torch.no_grad() def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): '''simple docstring''' if config_path is not None: A_ : Union[str, Any] = FlavaConfig.from_pretrained(lowercase__ ) else: A_ : Union[str, Any] = FlavaConfig() A_ : str = FlavaForPreTraining(lowercase__ ).eval() A_ : Union[str, Any] = convert_dalle_checkpoint(lowercase__ , lowercase__ , save_checkpoint=lowercase__ ) if os.path.exists(lowercase__ ): A_ : str = torch.load(lowercase__ , map_location="""cpu""" ) else: A_ : List[str] = torch.hub.load_state_dict_from_url(lowercase__ , map_location="""cpu""" ) A_ : Dict = upgrade_state_dict(lowercase__ , lowercase__ ) hf_model.load_state_dict(lowercase__ ) A_ : Union[str, Any] = hf_model.state_dict() A_ : Any = count_parameters(lowercase__ ) A_ : str = count_parameters(lowercase__ ) + count_parameters(lowercase__ ) assert torch.allclose(lowercase__ , lowercase__ , atol=1E-3 ) hf_model.save_pretrained(lowercase__ ) if __name__ == "__main__": lowerCamelCase :Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to flava checkpoint''') parser.add_argument('''--codebook_path''', default=None, type=str, help='''Path to flava codebook checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCamelCase :Optional[int] = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
351
'''simple docstring''' import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _lowerCAmelCase ( unittest.TestCase ): def _a (self ): A_ : Dict = 10 def _a (self ): A_ : str = [1, 2, 3, 4] A_ : List[str] = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(lowercase , self.block_size , 0 ) , lowercase ) def _a (self ): A_ : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] A_ : Tuple = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(lowercase , self.block_size , 0 ) , lowercase ) def _a (self ): A_ : List[Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] A_ : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(lowercase , self.block_size , 0 ) , lowercase ) def _a (self ): A_ : Any = """It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.""" A_, A_ : Optional[int] = process_story(lowercase ) self.assertEqual(lowercase , [] ) def _a (self ): A_ : Union[str, Any] = """""" A_, A_ : Union[str, Any] = process_story(lowercase ) self.assertEqual(lowercase , [] ) self.assertEqual(lowercase , [] ) def _a (self ): A_ : List[Any] = ( """It was the year of Our Lord one thousand seven hundred and """ """seventy-five\n\nSpiritual revelations were conceded to England """ """at that favoured period, as at this.\n@highlight\n\nIt was the best of times""" ) A_, A_ : int = process_story(lowercase ) A_ : List[Any] = [ """It was the year of Our Lord one thousand seven hundred and seventy-five.""", """Spiritual revelations were conceded to England at that favoured period, as at this.""", ] self.assertEqual(lowercase , lowercase ) A_ : Tuple = ["""It was the best of times."""] self.assertEqual(lowercase , lowercase ) def _a (self ): A_ : Dict = torch.tensor([1, 2, 3, 4] ) A_ : Union[str, Any] = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(lowercase , 0 ).numpy() , expected.numpy() ) def _a (self ): A_ : Union[str, Any] = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) A_ : List[str] = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowercase , 23 ).numpy() , expected.numpy() ) def _a (self ): A_ : Optional[Any] = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) A_ : Tuple = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowercase , 1 ).numpy() , expected.numpy() ) def _a (self ): A_ : List[str] = 101 A_ : str = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]] ) A_ : int = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) A_ : Optional[Any] = compute_token_type_ids(lowercase , lowercase ) np.testing.assert_array_equal(lowercase , lowercase )
135
0
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCamelCase : int , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=1024 , lowerCamelCase : Tuple=1024 , lowerCamelCase : Tuple=False , **lowerCamelCase : Any ): lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCamelCase ) lowerCAmelCase = SeqaSeqDataset(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , type_path='train' , **lowerCamelCase ) lowerCAmelCase = tok.pad_token_id def get_lens(lowerCamelCase : List[Any] ): lowerCAmelCase = tqdm( DataLoader(lowerCamelCase , batch_size=512 , num_workers=8 , shuffle=lowerCamelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) lowerCAmelCase = [] for batch in dl: lowerCAmelCase = batch['input_ids'].ne(lowerCamelCase ).sum(1 ).tolist() lowerCAmelCase = batch['labels'].ne(lowerCamelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCamelCase , lowerCamelCase ): max_lens.append(max(lowerCamelCase , lowerCamelCase ) ) else: max_lens.extend(lowerCamelCase ) return max_lens lowerCAmelCase = get_lens(lowerCamelCase ) lowerCAmelCase = SeqaSeqDataset(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , type_path='val' , **lowerCamelCase ) lowerCAmelCase = get_lens(lowerCamelCase ) pickle_save(lowerCamelCase , train_ds.len_file ) pickle_save(lowerCamelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
4
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class snake_case__ ( snake_case_ ): _snake_case : "DiagonalGaussianDistribution" class snake_case__ ( snake_case_, snake_case_ ): _snake_case : Optional[Any] = True @register_to_config def __init__( self , lowerCamelCase = 3 , lowerCamelCase = 3 , lowerCamelCase = ("DownEncoderBlock2D",) , lowerCamelCase = ("UpDecoderBlock2D",) , lowerCamelCase = (64,) , lowerCamelCase = 1 , lowerCamelCase = "silu" , lowerCamelCase = 4 , lowerCamelCase = 32 , lowerCamelCase = 32 , lowerCamelCase = 0.1_8215 , ): super().__init__() # pass init params to Encoder __a = Encoder( in_channels=lowerCamelCase , out_channels=lowerCamelCase , down_block_types=lowerCamelCase , block_out_channels=lowerCamelCase , layers_per_block=lowerCamelCase , act_fn=lowerCamelCase , norm_num_groups=lowerCamelCase , double_z=lowerCamelCase , ) # pass init params to Decoder __a = Decoder( in_channels=lowerCamelCase , out_channels=lowerCamelCase , up_block_types=lowerCamelCase , block_out_channels=lowerCamelCase , layers_per_block=lowerCamelCase , norm_num_groups=lowerCamelCase , act_fn=lowerCamelCase , ) __a = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) __a = nn.Convad(lowerCamelCase , lowerCamelCase , 1 ) __a = False __a = False # only relevant if vae tiling is enabled __a = self.config.sample_size __a = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) __a = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) __a = 0.25 def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if isinstance(lowerCamelCase , (Encoder, Decoder) ): __a = value def a__ ( self , lowerCamelCase = True ): __a = use_tiling def a__ ( self ): self.enable_tiling(lowerCamelCase ) def a__ ( self ): __a = True def a__ ( self ): __a = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def a__ ( self ): __a = {} def fn_recursive_add_processors(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , "set_processor" ): __a = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"{name}.{sub_name}" , lowerCamelCase , lowerCamelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return processors def a__ ( self , lowerCamelCase ): __a = len(self.attn_processors.keys() ) if isinstance(lowerCamelCase , lowerCamelCase ) and len(lowerCamelCase ) != count: raise ValueError( F"A dict of processors was passed, but the number of processors {len(lowerCamelCase )} does not match the" F" number of attention layers: {count}. Please make sure to pass {count} processor classes." ) def fn_recursive_attn_processor(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , "set_processor" ): if not isinstance(lowerCamelCase , lowerCamelCase ): module.set_processor(lowerCamelCase ) else: module.set_processor(processor.pop(F"{name}.processor" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"{name}.{sub_name}" , lowerCamelCase , lowerCamelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def a__ ( self ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(lowerCamelCase , return_dict=lowerCamelCase ) if self.use_slicing and x.shape[0] > 1: __a = [self.encoder(lowerCamelCase ) for x_slice in x.split(1 )] __a = torch.cat(lowerCamelCase ) else: __a = self.encoder(lowerCamelCase ) __a = self.quant_conv(lowerCamelCase ) __a = DiagonalGaussianDistribution(lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(lowerCamelCase , return_dict=lowerCamelCase ) __a = self.post_quant_conv(lowerCamelCase ) __a = self.decoder(lowerCamelCase ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase ) @apply_forward_hook def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_slicing and z.shape[0] > 1: __a = [self._decode(lowerCamelCase ).sample for z_slice in z.split(1 )] __a = torch.cat(lowerCamelCase ) else: __a = self._decode(lowerCamelCase ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = min(a.shape[2] , b.shape[2] , lowerCamelCase ) for y in range(lowerCamelCase ): __a = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = min(a.shape[3] , b.shape[3] , lowerCamelCase ) for x in range(lowerCamelCase ): __a = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def a__ ( self , lowerCamelCase , lowerCamelCase = True ): __a = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) __a = int(self.tile_latent_min_size * self.tile_overlap_factor ) __a = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. __a = [] for i in range(0 , x.shape[2] , lowerCamelCase ): __a = [] for j in range(0 , x.shape[3] , lowerCamelCase ): __a = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] __a = self.encoder(lowerCamelCase ) __a = self.quant_conv(lowerCamelCase ) row.append(lowerCamelCase ) rows.append(lowerCamelCase ) __a = [] for i, row in enumerate(lowerCamelCase ): __a = [] for j, tile in enumerate(lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __a = self.blend_v(rows[i - 1][j] , lowerCamelCase , lowerCamelCase ) if j > 0: __a = self.blend_h(row[j - 1] , lowerCamelCase , lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase , dim=3 ) ) __a = torch.cat(lowerCamelCase , dim=2 ) __a = DiagonalGaussianDistribution(lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = True ): __a = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) __a = int(self.tile_sample_min_size * self.tile_overlap_factor ) __a = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. __a = [] for i in range(0 , z.shape[2] , lowerCamelCase ): __a = [] for j in range(0 , z.shape[3] , lowerCamelCase ): __a = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] __a = self.post_quant_conv(lowerCamelCase ) __a = self.decoder(lowerCamelCase ) row.append(lowerCamelCase ) rows.append(lowerCamelCase ) __a = [] for i, row in enumerate(lowerCamelCase ): __a = [] for j, tile in enumerate(lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __a = self.blend_v(rows[i - 1][j] , lowerCamelCase , lowerCamelCase ) if j > 0: __a = self.blend_h(row[j - 1] , lowerCamelCase , lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase , dim=3 ) ) __a = torch.cat(lowerCamelCase , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = False , lowerCamelCase = True , lowerCamelCase = None , ): __a = sample __a = self.encode(lowerCamelCase ).latent_dist if sample_posterior: __a = posterior.sample(generator=lowerCamelCase ) else: __a = posterior.mode() __a = self.decode(lowerCamelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase )
261
0
import numpy as np def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1E-12 , lowerCAmelCase__ = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowerCAmelCase__ )[0] == np.shape(lowerCAmelCase__ )[1] # Ensure proper dimensionality. assert np.shape(lowerCAmelCase__ )[0] == np.shape(lowerCAmelCase__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowerCAmelCase__ ) == np.iscomplexobj(lowerCAmelCase__ ) UpperCAmelCase__ : str = np.iscomplexobj(lowerCAmelCase__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowerCAmelCase__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : Tuple = 1E12 while not convergence: # Multiple matrix by the vector. UpperCAmelCase__ : Optional[Any] = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) # Normalize the resulting output vector. UpperCAmelCase__ : Optional[int] = w / np.linalg.norm(lowerCAmelCase__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) UpperCAmelCase__ : str = vector.conj().T if is_complex else vector.T UpperCAmelCase__ : List[str] = np.dot(lowerCAmelCase__ , np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) ) # Check convergence. UpperCAmelCase__ : int = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : Dict = lambda_ if is_complex: UpperCAmelCase__ : List[Any] = np.real(lambda_ ) return lambda_, vector def a__ ( ) -> None: UpperCAmelCase__ : Tuple = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) UpperCAmelCase__ : str = np.array([41, 4, 20] ) UpperCAmelCase__ : List[Any] = real_input_matrix.astype(np.complexaaa ) UpperCAmelCase__ : List[str] = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T UpperCAmelCase__ : Dict = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": UpperCAmelCase__ : str = real_input_matrix UpperCAmelCase__ : Union[str, Any] = real_vector elif problem_type == "complex": UpperCAmelCase__ : Dict = complex_input_matrix UpperCAmelCase__ : Optional[int] = complex_vector # Our implementation. UpperCAmelCase__ : Union[str, Any] = power_iteration(lowerCAmelCase__ , lowerCAmelCase__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). UpperCAmelCase__ : Union[str, Any] = np.linalg.eigh(lowerCAmelCase__ ) # Last eigenvalue is the maximum one. UpperCAmelCase__ : List[str] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. UpperCAmelCase__ : Tuple = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1E-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowerCAmelCase__ ) - np.abs(lowerCAmelCase__ ) ) <= 1E-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
352
'''simple docstring''' class lowerCamelCase_ : def __init__( self : Union[str, Any] , _A : int ): '''simple docstring''' UpperCAmelCase__ : str = n UpperCAmelCase__ : Union[str, Any] = [None] * self.n UpperCAmelCase__ : Tuple = 0 # index of the first element UpperCAmelCase__ : int = 0 UpperCAmelCase__ : int = 0 def __len__( self : Optional[Any] ): '''simple docstring''' return self.size def lowercase_ ( self : Dict ): '''simple docstring''' return self.size == 0 def lowercase_ ( self : List[str] ): '''simple docstring''' return False if self.is_empty() else self.array[self.front] def lowercase_ ( self : List[Any] , _A : int ): '''simple docstring''' if self.size >= self.n: raise Exception('''QUEUE IS FULL''' ) UpperCAmelCase__ : str = data UpperCAmelCase__ : Optional[Any] = (self.rear + 1) % self.n self.size += 1 return self def lowercase_ ( self : List[Any] ): '''simple docstring''' if self.size == 0: raise Exception('''UNDERFLOW''' ) UpperCAmelCase__ : Any = self.array[self.front] UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Tuple = (self.front + 1) % self.n self.size -= 1 return temp
299
0
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = filter(lambda lowercase__ : p.requires_grad , model.parameters() ) __lowerCAmelCase : Any = sum([np.prod(p.size() ) for p in model_parameters] ) return params _UpperCamelCase = logging.getLogger(__name__) def _lowercase ( lowercase__ , lowercase__ ): if metric == "rouge2": __lowerCAmelCase : List[str] = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": __lowerCAmelCase : Union[str, Any] = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": __lowerCAmelCase : str = '''{val_avg_em:.4f}-{step_count}''' else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ''' function.''' ) __lowerCAmelCase : Union[str, Any] = ModelCheckpoint( dirpath=lowercase__ , filename=lowercase__ , monitor=f"""val_{metric}""" , mode='''max''' , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def _lowercase ( lowercase__ , lowercase__ ): return EarlyStopping( monitor=f"""val_{metric}""" , mode='''min''' if '''loss''' in metric else '''max''' , patience=lowercase__ , verbose=lowercase__ , ) class __lowercase (pl.Callback ): def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : Tuple = {f"""lr_group_{i}""": param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(A_ ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ , A_ , A_=True ) ->None: '''simple docstring''' logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase : List[Any] = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']} ) # Log results __lowerCAmelCase : int = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase : int = od / '''test_results.txt''' __lowerCAmelCase : Tuple = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase : List[Any] = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase : int = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=A_ ) generations_file.parent.mkdir(exist_ok=A_ ) with open(A_ , '''a+''' ) as writer: for key in sorted(A_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase : List[Any] = metrics[key] if isinstance(A_ , torch.Tensor ): __lowerCAmelCase : Union[str, Any] = val.item() __lowerCAmelCase : Dict = f"""{key}: {val:.6f}\n""" writer.write(A_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase : List[Any] = '''\n'''.join(metrics['''preds'''] ) generations_file.open('''w+''' ).write(A_ ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' try: __lowerCAmelCase : Optional[int] = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase : List[Any] = pl_module.model.num_parameters() __lowerCAmelCase : Optional[int] = count_trainable_parameters(A_ ) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1e6, '''grad_mp''': n_trainable_pars / 1e6} ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->List[Any]: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(A_ , A_ , '''test''' ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->List[str]: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
275
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
1
from math import factorial def a( A : int = 100 ) -> int: """simple docstring""" return sum(int(A ) for x in str(factorial(A ) ) ) if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
71
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( lowerCAmelCase, unittest.TestCase ): """simple docstring""" __A = LxmertTokenizer __A = LxmertTokenizerFast __A = True __A = True def UpperCamelCase_ (self ): """simple docstring""" super().setUp() a = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" a = "UNwant\u00E9d,running" a = "unwanted, running" return input_text, output_text def UpperCamelCase_ (self ): """simple docstring""" a = self.tokenizer_class(self.vocab_file ) a = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(lowerCamelCase_ , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [7, 4, 5, 10, 8, 9] ) def UpperCamelCase_ (self ): """simple docstring""" if not self.test_rust_tokenizer: return a = self.get_tokenizer() a = self.get_rust_tokenizer() a = "I was born in 92000, and this is falsé." a = tokenizer.tokenize(lowerCamelCase_ ) a = rust_tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) a = tokenizer.encode(lowerCamelCase_ , add_special_tokens=lowerCamelCase_ ) a = rust_tokenizer.encode(lowerCamelCase_ , add_special_tokens=lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) a = self.get_rust_tokenizer() a = tokenizer.encode(lowerCamelCase_ ) a = rust_tokenizer.encode(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ )
71
1
'''simple docstring''' def UpperCAmelCase_ ( __lowercase : int ) -> list: '''simple docstring''' _UpperCAmelCase = int(__lowercase ) if n_element < 1: _UpperCAmelCase = ValueError("a should be a positive number" ) raise my_error _UpperCAmelCase = [1] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = (0, 0, 0) _UpperCAmelCase = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": __SCREAMING_SNAKE_CASE :Any = input('''Enter the last number (nth term) of the Hamming Number Series: ''') print('''Formula of Hamming Number Series => 2^i * 3^j * 5^k''') __SCREAMING_SNAKE_CASE :Tuple = hamming(int(n)) print('''-----------------------------------------------------''') print(F"The list with nth numbers is: {hamming_numbers}") print('''-----------------------------------------------------''')
22
'''simple docstring''' from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent lowerCamelCase__ = {'UserAgent': UserAgent().random} def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[str] = script.contents[0] _UpperCAmelCase : Any = json.loads(data[data.find("{\"config\"" ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class lowerCAmelCase__ : def __init__( self : str , lowerCamelCase__ : List[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = F"""https://www.instagram.com/{username}/""" _UpperCAmelCase : Dict = self.get_json() def lowerCAmelCase__ ( self : Any ) ->dict: '''simple docstring''' _UpperCAmelCase : Optional[int] = requests.get(self.url , headers=lowerCamelCase__ ).text _UpperCAmelCase : Any = BeautifulSoup(lowerCamelCase__ , "html.parser" ).find_all("script" ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : Optional[Any] ) ->str: '''simple docstring''' return F"""{self.__class__.__name__}('{self.username}')""" def __str__( self : List[Any] ) ->str: '''simple docstring''' return F"""{self.fullname} ({self.username}) is {self.biography}""" @property def lowerCAmelCase__ ( self : Optional[int] ) ->str: '''simple docstring''' return self.user_data["username"] @property def lowerCAmelCase__ ( self : Optional[int] ) ->str: '''simple docstring''' return self.user_data["full_name"] @property def lowerCAmelCase__ ( self : Optional[int] ) ->str: '''simple docstring''' return self.user_data["biography"] @property def lowerCAmelCase__ ( self : int ) ->str: '''simple docstring''' return self.user_data["business_email"] @property def lowerCAmelCase__ ( self : str ) ->str: '''simple docstring''' return self.user_data["external_url"] @property def lowerCAmelCase__ ( self : Tuple ) ->int: '''simple docstring''' return self.user_data["edge_followed_by"]["count"] @property def lowerCAmelCase__ ( self : str ) ->int: '''simple docstring''' return self.user_data["edge_follow"]["count"] @property def lowerCAmelCase__ ( self : Any ) ->int: '''simple docstring''' return self.user_data["edge_owner_to_timeline_media"]["count"] @property def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' return self.user_data["profile_pic_url_hd"] @property def lowerCAmelCase__ ( self : Optional[Any] ) ->bool: '''simple docstring''' return self.user_data["is_verified"] @property def lowerCAmelCase__ ( self : int ) ->bool: '''simple docstring''' return self.user_data["is_private"] def __lowerCAmelCase (__lowerCAmelCase = "github" ): import os if os.environ.get("CI" ): return # test failing on GitHub Actions _UpperCAmelCase : Dict = InstagramUser(__lowerCAmelCase ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , __lowerCAmelCase ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 120_000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith("https://instagram." ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = InstagramUser('github') print(instagram_user) print(F'''{instagram_user.number_of_posts = }''') print(F'''{instagram_user.number_of_followers = }''') print(F'''{instagram_user.number_of_followings = }''') print(F'''{instagram_user.email = }''') print(F'''{instagram_user.website = }''') print(F'''{instagram_user.profile_picture_url = }''') print(F'''{instagram_user.is_verified = }''') print(F'''{instagram_user.is_private = }''')
234
0
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class _UpperCamelCase ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 class _UpperCamelCase ( nn.Module ): '''simple docstring''' lowerCAmelCase__ = 42 lowerCAmelCase__ = (16, 32, 96, 2_56) lowerCAmelCase__ = jnp.floataa def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase =nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) __lowercase =[] for i in range(len(self.block_out_channels) - 1): __lowercase =self.block_out_channels[i] __lowercase =self.block_out_channels[i + 1] __lowercase =nn.Conv( __snake_case , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(__snake_case) __lowercase =nn.Conv( __snake_case , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(__snake_case) __lowercase =blocks __lowercase =nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self : Tuple , _lowerCAmelCase : List[str]): '''simple docstring''' __lowercase =self.conv_in(__snake_case) __lowercase =nn.silu(__snake_case) for block in self.blocks: __lowercase =block(__snake_case) __lowercase =nn.silu(__snake_case) __lowercase =self.conv_out(__snake_case) return embedding @flax_register_to_config class _UpperCamelCase ( nn.Module , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase__ = 32 lowerCAmelCase__ = 4 lowerCAmelCase__ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowerCAmelCase__ = False lowerCAmelCase__ = (3_20, 6_40, 12_80, 12_80) lowerCAmelCase__ = 2 lowerCAmelCase__ = 8 lowerCAmelCase__ = None lowerCAmelCase__ = 12_80 lowerCAmelCase__ = 0.0 lowerCAmelCase__ = False lowerCAmelCase__ = jnp.floataa lowerCAmelCase__ = True lowerCAmelCase__ = 0 lowerCAmelCase__ = "rgb" lowerCAmelCase__ = (16, 32, 96, 2_56) def __lowerCamelCase ( self : Any , _lowerCAmelCase : jax.random.KeyArray): '''simple docstring''' __lowercase =(1, self.in_channels, self.sample_size, self.sample_size) __lowercase =jnp.zeros(__snake_case , dtype=jnp.floataa) __lowercase =jnp.ones((1,) , dtype=jnp.intaa) __lowercase =jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa) __lowercase =(1, 3, self.sample_size * 8, self.sample_size * 8) __lowercase =jnp.zeros(__snake_case , dtype=jnp.floataa) __lowercase , __lowercase =jax.random.split(__snake_case) __lowercase ={'params': params_rng, 'dropout': dropout_rng} return self.init(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case)["params"] def __lowerCamelCase ( self : Dict): '''simple docstring''' __lowercase =self.block_out_channels __lowercase =block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. __lowercase =self.num_attention_heads or self.attention_head_dim # input __lowercase =nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time __lowercase =FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift) __lowercase =FlaxTimestepEmbedding(__snake_case , dtype=self.dtype) __lowercase =FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) __lowercase =self.only_cross_attention if isinstance(__snake_case , __snake_case): __lowercase =(only_cross_attention,) * len(self.down_block_types) if isinstance(__snake_case , __snake_case): __lowercase =(num_attention_heads,) * len(self.down_block_types) # down __lowercase =[] __lowercase =[] __lowercase =block_out_channels[0] __lowercase =nn.Conv( __snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(__snake_case) for i, down_block_type in enumerate(self.down_block_types): __lowercase =output_channel __lowercase =block_out_channels[i] __lowercase =i == len(__snake_case) - 1 if down_block_type == "CrossAttnDownBlock2D": __lowercase =FlaxCrossAttnDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: __lowercase =FlaxDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(__snake_case) for _ in range(self.layers_per_block): __lowercase =nn.Conv( __snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(__snake_case) if not is_final_block: __lowercase =nn.Conv( __snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(__snake_case) __lowercase =down_blocks __lowercase =controlnet_down_blocks # mid __lowercase =block_out_channels[-1] __lowercase =FlaxUNetMidBlockaDCrossAttn( in_channels=__snake_case , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) __lowercase =nn.Conv( __snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : float = 1.0 , _lowerCAmelCase : bool = True , _lowerCAmelCase : bool = False , ): '''simple docstring''' __lowercase =self.controlnet_conditioning_channel_order if channel_order == "bgr": __lowercase =jnp.flip(__snake_case , axis=1) # 1. time if not isinstance(__snake_case , jnp.ndarray): __lowercase =jnp.array([timesteps] , dtype=jnp.intaa) elif isinstance(__snake_case , jnp.ndarray) and len(timesteps.shape) == 0: __lowercase =timesteps.astype(dtype=jnp.floataa) __lowercase =jnp.expand_dims(__snake_case , 0) __lowercase =self.time_proj(__snake_case) __lowercase =self.time_embedding(__snake_case) # 2. pre-process __lowercase =jnp.transpose(__snake_case , (0, 2, 3, 1)) __lowercase =self.conv_in(__snake_case) __lowercase =jnp.transpose(__snake_case , (0, 2, 3, 1)) __lowercase =self.controlnet_cond_embedding(__snake_case) sample += controlnet_cond # 3. down __lowercase =(sample,) for down_block in self.down_blocks: if isinstance(__snake_case , __snake_case): __lowercase , __lowercase =down_block(__snake_case , __snake_case , __snake_case , deterministic=not train) else: __lowercase , __lowercase =down_block(__snake_case , __snake_case , deterministic=not train) down_block_res_samples += res_samples # 4. mid __lowercase =self.mid_block(__snake_case , __snake_case , __snake_case , deterministic=not train) # 5. contronet blocks __lowercase =() for down_block_res_sample, controlnet_block in zip(__snake_case , self.controlnet_down_blocks): __lowercase =controlnet_block(__snake_case) controlnet_down_block_res_samples += (down_block_res_sample,) __lowercase =controlnet_down_block_res_samples __lowercase =self.controlnet_mid_block(__snake_case) # 6. scaling __lowercase =[sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=__snake_case , mid_block_res_sample=__snake_case)
370
'''simple docstring''' import functools def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =len(_lowerCAmelCase ) __lowercase =len(_lowerCAmelCase ) @functools.cache def min_distance(_lowerCAmelCase , _lowerCAmelCase ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa __lowercase =int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , _lowerCAmelCase ) , 1 + min_distance(_lowerCAmelCase , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
48
0
"""simple docstring""" import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class _snake_case ( unittest.TestCase ): def lowerCamelCase__ ( self : Any ): # A mock response for an HTTP head request to emulate server down __lowerCamelCase : Optional[Any] = mock.Mock() __lowerCamelCase : Union[str, Any] = 500 __lowerCamelCase : Optional[Any] = {} __lowerCamelCase : int = HTTPError __lowerCamelCase : Optional[int] = {} # Download this model to make sure it's in the cache. __lowerCamelCase : List[Any] = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=UpperCAmelCase ) as mock_head: __lowerCamelCase : str = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def lowerCamelCase__ ( self : Tuple ): # A mock response for an HTTP head request to emulate server down __lowerCamelCase : Dict = mock.Mock() __lowerCamelCase : Optional[Any] = 500 __lowerCamelCase : int = {} __lowerCamelCase : Tuple = HTTPError __lowerCamelCase : List[Any] = {} # Download this model to make sure it's in the cache. __lowerCamelCase : Optional[int] = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=UpperCAmelCase ) as mock_head: __lowerCamelCase : Union[str, Any] = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def lowerCamelCase__ ( self : Any ): # This test is for deprecated behavior and can be removed in v5 try: __lowerCamelCase : Tuple = tempfile.mktemp() with open(UpperCAmelCase , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , UpperCAmelCase ) __lowerCamelCase : Dict = AlbertTokenizer.from_pretrained(UpperCAmelCase ) finally: os.remove(UpperCAmelCase ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , UpperCAmelCase ) __lowerCamelCase : int = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def lowerCamelCase__ ( self : List[Any] ): # This test is for deprecated behavior and can be removed in v5 __lowerCamelCase : Optional[int] = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class _snake_case ( unittest.TestCase ): snake_case__ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def lowerCamelCase__ ( cls : List[str] ): __lowerCamelCase : Tuple = TOKEN HfFolder.save_token(UpperCAmelCase ) @classmethod def lowerCamelCase__ ( cls : List[str] ): try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def lowerCamelCase__ ( self : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : str = os.path.join(UpperCAmelCase , "vocab.txt" ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __lowerCamelCase : int = BertTokenizer(UpperCAmelCase ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __lowerCamelCase : int = BertTokenizer.from_pretrained(F"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase , repo_id="test-tokenizer" , push_to_hub=UpperCAmelCase , use_auth_token=self._token ) __lowerCamelCase : int = BertTokenizer.from_pretrained(F"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def lowerCamelCase__ ( self : int ): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Tuple = os.path.join(UpperCAmelCase , "vocab.txt" ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __lowerCamelCase : Optional[int] = BertTokenizer(UpperCAmelCase ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __lowerCamelCase : Tuple = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( UpperCAmelCase , repo_id="valid_org/test-tokenizer-org" , push_to_hub=UpperCAmelCase , use_auth_token=self._token ) __lowerCamelCase : Optional[int] = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def lowerCamelCase__ ( self : Any ): CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Optional[Any] = os.path.join(UpperCAmelCase , "vocab.txt" ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __lowerCamelCase : str = CustomTokenizer(UpperCAmelCase ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __lowerCamelCase : List[str] = AutoTokenizer.from_pretrained(F"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Tuple = os.path.join(UpperCAmelCase , "vocab.txt" ) with open(UpperCAmelCase , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __lowerCamelCase : Optional[Any] = BertTokenizerFast.from_pretrained(UpperCAmelCase ) bert_tokenizer.save_pretrained(UpperCAmelCase ) __lowerCamelCase : int = CustomTokenizerFast.from_pretrained(UpperCAmelCase ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __lowerCamelCase : int = AutoTokenizer.from_pretrained(F"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained( F"""{USER}/test-dynamic-tokenizer""" , use_fast=UpperCAmelCase , trust_remote_code=UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class _snake_case ( unittest.TestCase ): def lowerCamelCase__ ( self : Tuple ): __lowerCamelCase : Optional[int] = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def lowerCamelCase__ ( self : List[Any] ): __lowerCamelCase : int = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def lowerCamelCase__ ( self : List[Any] ): __lowerCamelCase : Optional[int] = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def lowerCamelCase__ ( self : Optional[int] ): __lowerCamelCase : Optional[Any] = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def lowerCamelCase__ ( self : Tuple ): __lowerCamelCase : Optional[Any] = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def lowerCamelCase__ ( self : Dict ): __lowerCamelCase : List[Any] = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def lowerCamelCase__ ( self : Any ): __lowerCamelCase : Union[str, Any] = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def lowerCamelCase__ ( self : List[str] ): # Even if the offsets are wrong, we necessarily output correct string # parts. __lowerCamelCase : Optional[int] = Trie() __lowerCamelCase : str = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(UpperCAmelCase , ["AB", "C"] )
135
"""simple docstring""" def lowercase_ ( _lowerCamelCase: Dict ) -> List[str]: '''simple docstring''' __lowerCamelCase : Tuple = 1 __lowerCamelCase : int = 2 while i * i <= n: __lowerCamelCase : Tuple = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def lowercase_ ( ) -> str: '''simple docstring''' __lowerCamelCase : List[str] = 1 __lowerCamelCase : Dict = 1 while True: i += 1 t_num += i if count_divisors(_lowerCamelCase ) > 500: break return t_num if __name__ == "__main__": print(solution())
135
1
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __A : def __init__( self , a__ = "cpu" , a__ = "openai/clip-vit-large-patch14" ): _lowerCAmelCase : Optional[int] = device _lowerCAmelCase : Tuple = CLIPTokenizerFast.from_pretrained(a__ ) _lowerCAmelCase : str = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] _lowerCAmelCase : List[Any] = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] _lowerCAmelCase : Any = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _lowerCAmelCase : Union[str, Any] = torchvision.transforms.Resize(224 ) _lowerCAmelCase : Dict = torchvision.transforms.CenterCrop(224 ) def __A ( self , a__ ): _lowerCAmelCase : str = self.resize(a__ ) _lowerCAmelCase : Optional[Any] = self.center_crop(a__ ) _lowerCAmelCase : List[str] = self.normalize(a__ ) return images def __call__( self , a__=None , a__=None , **a__ ): _lowerCAmelCase : Optional[Any] = self.tokenizer(text=a__ , **a__ ) _lowerCAmelCase : str = self.preprocess_img(a__ ) _lowerCAmelCase : Tuple = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class __A ( nn.Module ): def __init__( self , a__=10 , a__=0.0_1 , a__=None , a__=None , a__=None , a__=None , a__=None , a__=None , a__=False , a__=True , a__="image" , a__=True , a__=False , a__=False , a__=False , ): super().__init__() _lowerCAmelCase : List[str] = None _lowerCAmelCase : Optional[Any] = device if device else get_device() if vqgan: _lowerCAmelCase : Optional[int] = vqgan else: _lowerCAmelCase : Optional[Any] = load_vqgan(self.device , conf_path=a__ , ckpt_path=a__ ) self.vqgan.eval() if clip: _lowerCAmelCase : Tuple = clip else: _lowerCAmelCase : List[Any] = CLIPModel.from_pretrained("""openai/clip-vit-base-patch32""" ) self.clip.to(self.device ) _lowerCAmelCase : Optional[int] = ProcessorGradientFlow(device=self.device ) _lowerCAmelCase : Dict = iterations _lowerCAmelCase : int = lr _lowerCAmelCase : Tuple = log _lowerCAmelCase : List[str] = make_grid _lowerCAmelCase : List[Any] = return_val _lowerCAmelCase : Any = quantize _lowerCAmelCase : Optional[Any] = self.vqgan.decoder.z_shape def __A ( self , a__=None , a__=None , a__=5 , a__=True ): _lowerCAmelCase : List[Any] = [] if output_path is None: _lowerCAmelCase : Any = """./animation.gif""" if input_path is None: _lowerCAmelCase : int = self.save_path _lowerCAmelCase : Any = sorted(glob(input_path + """/*""" ) ) if not len(a__ ): raise ValueError( """No images found in save path, aborting (did you pass save_intermediate=True to the generate""" """ function?)""" ) if len(a__ ) == 1: print("""Only one image found in save path, (did you pass save_intermediate=True to the generate function?)""" ) _lowerCAmelCase : str = total_duration / len(a__ ) _lowerCAmelCase : int = [frame_duration] * len(a__ ) if extend_frames: _lowerCAmelCase : List[str] = 1.5 _lowerCAmelCase : List[Any] = 3 for file_name in paths: if file_name.endswith(""".png""" ): images.append(imageio.imread(a__ ) ) imageio.mimsave(a__ , a__ , duration=a__ ) print(F"gif saved to {output_path}" ) def __A ( self , a__=None , a__=None ): if not (path or img): raise ValueError("""Input either path or tensor""" ) if img is not None: raise NotImplementedError _lowerCAmelCase : Optional[Any] = preprocess(Image.open(a__ ) , target_image_size=256 ).to(self.device ) _lowerCAmelCase : List[str] = preprocess_vqgan(a__ ) _lowerCAmelCase , *_lowerCAmelCase : str = self.vqgan.encode(a__ ) return z def __A ( self , a__ ): _lowerCAmelCase : Tuple = self.latent.detach().requires_grad_() _lowerCAmelCase : Optional[Any] = base_latent + transform_vector if self.quantize: _lowerCAmelCase , *_lowerCAmelCase : Union[str, Any] = self.vqgan.quantize(a__ ) else: _lowerCAmelCase : List[str] = trans_latent return self.vqgan.decode(a__ ) def __A ( self , a__ , a__ , a__=None ): _lowerCAmelCase : Dict = self.clip_preprocessor(text=a__ , images=a__ , return_tensors="""pt""" , padding=a__ ) _lowerCAmelCase : List[Any] = self.clip(**a__ ) _lowerCAmelCase : Optional[Any] = clip_outputs.logits_per_image if weights is not None: _lowerCAmelCase : Union[str, Any] = similarity_logits * weights return similarity_logits.sum() def __A ( self , a__ , a__ , a__ ): _lowerCAmelCase : Tuple = self._get_clip_similarity(pos_prompts["""prompts"""] , a__ , weights=(1 / pos_prompts["""weights"""]) ) if neg_prompts: _lowerCAmelCase : Any = self._get_clip_similarity(neg_prompts["""prompts"""] , a__ , weights=neg_prompts["""weights"""] ) else: _lowerCAmelCase : List[str] = torch.tensor([1] , device=self.device ) _lowerCAmelCase : Any = -torch.log(a__ ) + torch.log(a__ ) return loss def __A ( self , a__ , a__ , a__ ): _lowerCAmelCase : Union[str, Any] = torch.randn_like(self.latent , requires_grad=a__ , device=self.device ) _lowerCAmelCase : Optional[Any] = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _lowerCAmelCase : Dict = self._add_vector(a__ ) _lowerCAmelCase : Any = loop_post_process(a__ ) _lowerCAmelCase : Dict = self._get_CLIP_loss(a__ , a__ , a__ ) print("""CLIP loss""" , a__ ) if self.log: wandb.log({"""CLIP Loss""": clip_loss} ) clip_loss.backward(retain_graph=a__ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def __A ( self , a__ , a__ , a__ ): wandb.init(reinit=a__ , project="""face-editor""" ) wandb.config.update({"""Positive Prompts""": positive_prompts} ) wandb.config.update({"""Negative Prompts""": negative_prompts} ) wandb.config.update({"""lr""": self.lr, """iterations""": self.iterations} ) if image_path: _lowerCAmelCase : List[str] = Image.open(a__ ) _lowerCAmelCase : int = image.resize((256, 256) ) wandb.log("""Original Image""" , wandb.Image(a__ ) ) def __A ( self , a__ ): if not prompts: return [] _lowerCAmelCase : Dict = [] _lowerCAmelCase : str = [] if isinstance(a__ , a__ ): _lowerCAmelCase : Dict = [prompt.strip() for prompt in prompts.split("""|""" )] for prompt in prompts: if isinstance(a__ , (tuple, list) ): _lowerCAmelCase : List[str] = prompt[0] _lowerCAmelCase : List[Any] = float(prompt[1] ) elif ":" in prompt: _lowerCAmelCase , _lowerCAmelCase : Optional[int] = prompt.split(""":""" ) _lowerCAmelCase : Tuple = float(a__ ) else: _lowerCAmelCase : List[str] = prompt _lowerCAmelCase : Dict = 1.0 processed_prompts.append(a__ ) weights.append(a__ ) return { "prompts": processed_prompts, "weights": torch.tensor(a__ , device=self.device ), } def __A ( self , a__ , a__=None , a__=None , a__=True , a__=False , a__=True , a__=True , a__=None , ): if image_path: _lowerCAmelCase : List[Any] = self._get_latent(a__ ) else: _lowerCAmelCase : Optional[int] = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(a__ , a__ , a__ ) assert pos_prompts, "You must provide at least one positive prompt." _lowerCAmelCase : Any = self.process_prompts(a__ ) _lowerCAmelCase : Optional[Any] = self.process_prompts(a__ ) if save_final and save_path is None: _lowerCAmelCase : Optional[Any] = os.path.join("""./outputs/""" , """_""".join(pos_prompts["""prompts"""] ) ) if not os.path.exists(a__ ): os.makedirs(a__ ) else: _lowerCAmelCase : str = save_path + """_""" + get_timestamp() os.makedirs(a__ ) _lowerCAmelCase : Optional[int] = save_path _lowerCAmelCase : Optional[int] = self.vqgan.decode(self.latent )[0] if show_intermediate: print("""Original Image""" ) show_pil(custom_to_pil(a__ ) ) _lowerCAmelCase : List[str] = loop_post_process(a__ ) for iter, transformed_img in enumerate(self._optimize_CLIP(a__ , a__ , a__ ) ): if show_intermediate: show_pil(a__ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}.png" ) ) if self.log: wandb.log({"""Image""": wandb.Image(a__ )} ) if show_final: show_pil(a__ ) if save_final: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}_final.png" ) )
126
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _a : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _a : Union[str, Any] = 250_004 _a : Optional[int] = 250_020 @require_sentencepiece @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Optional[Any] = MBartaaTokenizer _UpperCamelCase : Any = MBartaaTokenizerFast _UpperCamelCase : List[str] = True _UpperCamelCase : Optional[int] = True def __A ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : Tuple = MBartaaTokenizer(a__ , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=a__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : Any = """<s>""" _lowerCAmelCase : Optional[int] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a__ ) , a__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a__ ) , a__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(a__ ) , 1054 ) def __A ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1054 ) def __A ( self ): _lowerCAmelCase : str = MBartaaTokenizer(a__ , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=a__ ) _lowerCAmelCase : Optional[int] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : int = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a__ , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """."""] , ) _lowerCAmelCase : Union[str, Any] = tokenizer.convert_tokens_to_ids(a__ ) self.assertListEqual( a__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : int = tokenizer.convert_ids_to_tokens(a__ ) self.assertListEqual( a__ , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """."""] , ) @slow def __A ( self ): # fmt: off _lowerCAmelCase : Any = {"""input_ids""": [[250004, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [250004, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [250004, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a__ , model_name="""facebook/mbart-large-50""" , revision="""d3913889c59cd5c9e456b269c376325eabad57e2""" , ) def __A ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCAmelCase : List[str] = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-mbart50""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): _lowerCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(a__ , **a__ ) _lowerCAmelCase : List[str] = self.tokenizer_class.from_pretrained(a__ , **a__ ) _lowerCAmelCase : Optional[int] = tempfile.mkdtemp() _lowerCAmelCase : Optional[Any] = tokenizer_r.save_pretrained(a__ ) _lowerCAmelCase : List[str] = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) _lowerCAmelCase : Any = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : List[Any] = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : int = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=True _lowerCAmelCase : List[Any] = tempfile.mkdtemp() _lowerCAmelCase : Dict = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) _lowerCAmelCase : str = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : List[str] = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : Any = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=False _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Tuple = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) _lowerCAmelCase : str = tokenizer_p.save_pretrained(a__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCAmelCase : Any = tokenizer_r.from_pretrained(a__ ) _lowerCAmelCase : int = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) @require_torch @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): _UpperCamelCase : Union[str, Any] = "facebook/mbart-large-50-one-to-many-mmt" _UpperCamelCase : int = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] _UpperCamelCase : Any = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] _UpperCamelCase : List[str] = [EN_CODE, 8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2] @classmethod def __A ( cls ): _lowerCAmelCase : MBartaaTokenizer = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en_XX""" , tgt_lang="""ro_RO""" ) _lowerCAmelCase : str = 1 return cls def __A ( self ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ar_AR"""] , 250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""en_EN"""] , 250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ro_RO"""] , 250020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""mr_IN"""] , 250038 ) def __A ( self ): _lowerCAmelCase : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , a__ ) def __A ( self ): self.assertIn(a__ , self.tokenizer.all_special_ids ) _lowerCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] _lowerCAmelCase : int = self.tokenizer.decode(a__ , skip_special_tokens=a__ ) _lowerCAmelCase : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=a__ ) self.assertEqual(a__ , a__ ) self.assertNotIn(self.tokenizer.eos_token , a__ ) def __A ( self ): _lowerCAmelCase : Any = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , a__ ) _lowerCAmelCase : Optional[Any] = 10 _lowerCAmelCase : Optional[int] = self.tokenizer(a__ , max_length=a__ , truncation=a__ ).input_ids[0] self.assertEqual(ids[0] , a__ ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(a__ ) , a__ ) def __A ( self ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [250053, 250001] ) def __A ( self ): _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Tuple = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(a__ ) _lowerCAmelCase : List[Any] = MBartaaTokenizer.from_pretrained(a__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , a__ ) @require_torch def __A ( self ): _lowerCAmelCase : List[str] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=a__ , return_tensors="""pt""" ) _lowerCAmelCase : Any = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def __A ( self ): _lowerCAmelCase : Optional[int] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) _lowerCAmelCase : Tuple = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) self.assertIsInstance(a__ , a__ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) _lowerCAmelCase : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , a__ ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def __A ( self ): _lowerCAmelCase : str = self.tokenizer(self.src_text , padding=a__ , truncation=a__ , max_length=3 , return_tensors="""pt""" ) _lowerCAmelCase : List[Any] = self.tokenizer( text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=10 , return_tensors="""pt""" ) _lowerCAmelCase : List[str] = targets["""input_ids"""] _lowerCAmelCase : Any = shift_tokens_right(a__ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def __A ( self ): _lowerCAmelCase : str = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""en_XX""" , tgt_lang="""ar_AR""" ) self.assertEqual( nested_simplify(a__ ) , { # en_XX, A, test, EOS """input_ids""": [[250004, 62, 3034, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 250001, } , )
126
1
def a ( A__ : Tuple ) -> List[Any]: """simple docstring""" return " ".join( ''.join(word[::-1] ) if len(__lowerCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('Hey wollef sroirraw'))
205
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" @register_to_config def __init__( self , _A , _A , _A , _A , _A , _A , _A , _A , _A , _A = False , ) -> List[str]: super().__init__() SCREAMING_SNAKE_CASE_ = nn.Embedding(_A , _A ) SCREAMING_SNAKE_CASE_ = nn.Embedding(_A , _A ) SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = nn.Dropout(p=_A ) SCREAMING_SNAKE_CASE_ = TaConfig( vocab_size=_A , d_model=_A , num_heads=_A , d_kv=_A , d_ff=_A , dropout_rate=_A , feed_forward_proj=_A , is_decoder=_A , is_encoder_decoder=_A , ) SCREAMING_SNAKE_CASE_ = nn.ModuleList() for lyr_num in range(_A ): SCREAMING_SNAKE_CASE_ = TaBlock(_A ) self.encoders.append(_A ) SCREAMING_SNAKE_CASE_ = TaLayerNorm(_A ) SCREAMING_SNAKE_CASE_ = nn.Dropout(p=_A ) def _UpperCamelCase ( self , _A , _A ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.token_embedder(_A ) SCREAMING_SNAKE_CASE_ = encoder_input_tokens.shape[1] SCREAMING_SNAKE_CASE_ = torch.arange(_A , device=encoder_input_tokens.device ) x += self.position_encoding(_A ) SCREAMING_SNAKE_CASE_ = self.dropout_pre(_A ) # inverted the attention mask SCREAMING_SNAKE_CASE_ = encoder_input_tokens.size() SCREAMING_SNAKE_CASE_ = self.get_extended_attention_mask(_A , _A ) for lyr in self.encoders: SCREAMING_SNAKE_CASE_ = lyr(_A , _A )[0] SCREAMING_SNAKE_CASE_ = self.layer_norm(_A ) return self.dropout_post(_A ), encoder_inputs_mask
299
0
"""simple docstring""" import socket def lowerCAmelCase_( ) -> Optional[Any]: _lowerCamelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase = socket.gethostname() _lowerCamelCase = 1_23_12 sock.connect((host, port) ) sock.send(B'''Hello server!''' ) with open('''Received_file''' , '''wb''' ) as out_file: print('''File opened''' ) print('''Receiving data...''' ) while True: _lowerCamelCase = sock.recv(10_24 ) if not data: break out_file.write(lowercase_ ) print('''Successfully received the file''' ) sock.close() print('''Connection closed''' ) if __name__ == "__main__": main()
73
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'mgp-str' def __init__( self , lowerCamelCase__=[3_2, 1_2_8] , lowerCamelCase__=4 , lowerCamelCase__=3 , lowerCamelCase__=2_7 , lowerCamelCase__=3_8 , lowerCamelCase__=5_0_2_5_7 , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=4.0 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=False , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = max_token_length _lowerCamelCase = num_character_labels _lowerCamelCase = num_bpe_labels _lowerCamelCase = num_wordpiece_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = mlp_ratio _lowerCamelCase = distilled _lowerCamelCase = layer_norm_eps _lowerCamelCase = drop_rate _lowerCamelCase = qkv_bias _lowerCamelCase = attn_drop_rate _lowerCamelCase = drop_path_rate _lowerCamelCase = output_aa_attentions _lowerCamelCase = initializer_range
73
1
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A ( a_ ) -> List[Any]: # picklable for multiprocessing return x.sum() def A ( a_ ) -> Tuple: # picklable for multiprocessing return i + 1 @dataclass class __A : """simple docstring""" UpperCamelCase__ : int UpperCamelCase__ : str class __A ( a ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ={} __UpperCamelCase : Union[str, Any] =[] __UpperCamelCase : List[str] =1 __UpperCamelCase : List[Any] =[1, 2] __UpperCamelCase : Dict ={'a': 1, 'b': 2} __UpperCamelCase : Union[str, Any] ={'a': [1, 2], 'b': [3, 4]} __UpperCamelCase : Tuple ={'a': {'1': 1}, 'b': 2} __UpperCamelCase : Union[str, Any] ={'a': 1, 'b': 2, 'c': 3, 'd': 4} __UpperCamelCase : Dict ={} __UpperCamelCase : Optional[Any] =[] __UpperCamelCase : str =2 __UpperCamelCase : str =[2, 3] __UpperCamelCase : Any ={'a': 2, 'b': 3} __UpperCamelCase : Any ={'a': [2, 3], 'b': [4, 5]} __UpperCamelCase : List[str] ={'a': {'1': 2}, 'b': 3} __UpperCamelCase : str ={'a': 2, 'b': 3, 'c': 4, 'd': 5} self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) __UpperCamelCase : Tuple =2 self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) __UpperCamelCase : Tuple ={'a': np.eye(2 ), 'b': np.zeros(3 ), 'c': np.ones(2 )} __UpperCamelCase : List[Any] ={'a': 2, 'b': 0, 'c': 2} __UpperCamelCase : Tuple ={ 'a': np.eye(2 ).astype(lowerCamelCase__ ), 'b': np.zeros(3 ).astype(lowerCamelCase__ ), 'c': np.ones(2 ).astype(lowerCamelCase__ ), } self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , map_numpy=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowerCamelCase__ , lowerCamelCase__ , map_numpy=lowerCamelCase__ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(lowerCamelCase__ , lowerCamelCase__ , map_numpy=lowerCamelCase__ , num_proc=lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowerCamelCase__ , lowerCamelCase__ , map_numpy=lowerCamelCase__ , num_proc=lowerCamelCase__ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(lowerCamelCase__ ): # can't pickle a local lambda map_nested(lambda lowerCamelCase__ : x + 1 , lowerCamelCase__ , num_proc=lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ={'a': 1, 'b': 2} __UpperCamelCase : Any ={'a': 3, 'b': 4} __UpperCamelCase : Union[str, Any] ={'a': 5, 'b': 6} __UpperCamelCase : Dict =sorted([('a', (1, 3, 5)), ('b', (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) ) , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" class __A : """simple docstring""" UpperCamelCase__ : int ="""bar""" __UpperCamelCase : List[Any] =Foo() self.assertEqual(foo.my_attr , 'bar' ) with temporary_assignment(lowerCamelCase__ , 'my_attr' , 'BAR' ): self.assertEqual(foo.my_attr , 'BAR' ) self.assertEqual(foo.my_attr , 'bar' ) @pytest.mark.parametrize( 'iterable_length, num_proc, expected_num_proc' ,[ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] ,) def A ( a_ ,a_ ,a_ ) -> List[Any]: with patch('datasets.utils.py_utils._single_map_nested' ) as mock_single_map_nested, patch( 'datasets.parallel.parallel.Pool' ) as mock_multiprocessing_pool: __UpperCamelCase : Union[str, Any] ={F'{i}': i for i in range(a_ )} __UpperCamelCase : List[str] =map_nested(lambda a_ : x + 10 ,a_ ,num_proc=a_ ,parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class __A ( a ): """simple docstring""" @require_tf def __lowercase ( self ): """simple docstring""" import tensorflow as tf from tensorflow.keras import layers __UpperCamelCase : Dict =layers.Dense(2 ) def gen_random_output(): __UpperCamelCase : Any =tf.random.uniform((1, 3) ) return model(lowerCamelCase__ ).numpy() with temp_seed(42 , set_tensorflow=lowerCamelCase__ ): __UpperCamelCase : Optional[int] =gen_random_output() with temp_seed(42 , set_tensorflow=lowerCamelCase__ ): __UpperCamelCase : Tuple =gen_random_output() __UpperCamelCase : List[str] =gen_random_output() np.testing.assert_equal(lowerCamelCase__ , lowerCamelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __lowercase ( self ): """simple docstring""" import torch def gen_random_output(): __UpperCamelCase : Tuple =torch.nn.Linear(3 , 2 ) __UpperCamelCase : List[str] =torch.rand(1 , 3 ) return model(lowerCamelCase__ ).detach().numpy() with temp_seed(42 , set_pytorch=lowerCamelCase__ ): __UpperCamelCase : Tuple =gen_random_output() with temp_seed(42 , set_pytorch=lowerCamelCase__ ): __UpperCamelCase : int =gen_random_output() __UpperCamelCase : Union[str, Any] =gen_random_output() np.testing.assert_equal(lowerCamelCase__ , lowerCamelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __lowercase ( self ): """simple docstring""" def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): __UpperCamelCase : List[str] =gen_random_output() with temp_seed(42 ): __UpperCamelCase : List[Any] =gen_random_output() __UpperCamelCase : List[str] =gen_random_output() np.testing.assert_equal(lowerCamelCase__ , lowerCamelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize('input_data' ,[{}] ) def A ( a_ ) -> Optional[Any]: __UpperCamelCase : int =NestedDataStructure(a_ ).data assert output_data == input_data @pytest.mark.parametrize( 'data, expected_output' ,[ ({}, []), ([], []), ('foo', ['foo']), (['foo', 'bar'], ['foo', 'bar']), ([['foo', 'bar']], ['foo', 'bar']), ([[['foo'], ['bar']]], ['foo', 'bar']), ([[['foo'], 'bar']], ['foo', 'bar']), ({'a': 1, 'b': 2}, [1, 2]), ({'a': [1, 2], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[[3], [4]]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, [4]]}, [1, 2, 3, 4]), ({'a': {'1': 1}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': [2]}, [1, 2]), ] ,) def A ( a_ ,a_ ) -> Optional[int]: __UpperCamelCase : List[str] =NestedDataStructure(a_ ).flatten() assert output == expected_output def A ( ) -> int: __UpperCamelCase : int =A(x=1 ,y='foobar' ) __UpperCamelCase : Dict ={'x': 1, 'y': 'foobar'} assert asdict(a_ ) == expected_output __UpperCamelCase : Tuple ={'a': {'b': A(x=10 ,y='foo' )}, 'c': [A(x=20 ,y='bar' )]} __UpperCamelCase : Optional[Any] ={'a': {'b': {'x': 10, 'y': 'foo'}}, 'c': [{'x': 20, 'y': 'bar'}]} assert asdict(a_ ) == expected_output with pytest.raises(a_ ): asdict([1, A(x=10 ,y='foo' )] ) def A ( a_ ) -> int: return text.split() def A ( a_ ) -> Optional[Any]: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A ( ) -> List[Any]: with Pool(2 ) as pool: __UpperCamelCase : Optional[Any] =list(iflatmap_unordered(a_ ,_split_text ,kwargs_iterable=[{'text': 'hello there'}] * 10 ) ) assert out.count('hello' ) == 10 assert out.count('there' ) == 10 assert len(a_ ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: __UpperCamelCase : Dict =list(iflatmap_unordered(a_ ,_split_text ,kwargs_iterable=[{'text': 'hello there'}] * 10 ) ) assert out.count('hello' ) == 10 assert out.count('there' ) == 10 assert len(a_ ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: __UpperCamelCase : List[Any] =[] for yield_time, content in iflatmap_unordered( a_ ,_aseconds_generator_of_aitems_with_timing ,kwargs_iterable=[{'content': 'a'}, {'content': 'b'}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(a_ ) assert out.count('a' ) == 2 assert out.count('b' ) == 2 assert len(a_ ) == 4
71
def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square(a_ ,a_ ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __UpperCamelCase : Optional[int] =update_area_of_max_square(a_ ,col + 1 ) __UpperCamelCase : List[str] =update_area_of_max_square(row + 1 ,col + 1 ) __UpperCamelCase : List[Any] =update_area_of_max_square(row + 1 ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : Dict =max(largest_square_area[0] ,a_ ) return sub_problem_sol else: return 0 __UpperCamelCase : Union[str, Any] =[0] update_area_of_max_square(0 ,0 ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square_using_dp_array( a_ ,a_ ,a_ ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __UpperCamelCase : Tuple =update_area_of_max_square_using_dp_array(a_ ,col + 1 ,a_ ) __UpperCamelCase : Optional[int] =update_area_of_max_square_using_dp_array(row + 1 ,col + 1 ,a_ ) __UpperCamelCase : Any =update_area_of_max_square_using_dp_array(row + 1 ,a_ ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : str =max(largest_square_area[0] ,a_ ) __UpperCamelCase : Any =sub_problem_sol return sub_problem_sol else: return 0 __UpperCamelCase : Tuple =[0] __UpperCamelCase : List[Any] =[[-1] * cols for _ in range(a_ )] update_area_of_max_square_using_dp_array(0 ,0 ,a_ ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Dict =[[0] * (cols + 1) for _ in range(rows + 1 )] __UpperCamelCase : int =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Optional[Any] =dp_array[row][col + 1] __UpperCamelCase : int =dp_array[row + 1][col + 1] __UpperCamelCase : Tuple =dp_array[row + 1][col] if mat[row][col] == 1: __UpperCamelCase : Tuple =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Any =max(dp_array[row][col] ,a_ ) else: __UpperCamelCase : Dict =0 return largest_square_area def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Any =[0] * (cols + 1) __UpperCamelCase : List[Any] =[0] * (cols + 1) __UpperCamelCase : Tuple =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Any =current_row[col + 1] __UpperCamelCase : Optional[Any] =next_row[col + 1] __UpperCamelCase : Union[str, Any] =next_row[col] if mat[row][col] == 1: __UpperCamelCase : Any =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Optional[int] =max(current_row[col] ,a_ ) else: __UpperCamelCase : List[str] =0 __UpperCamelCase : Optional[Any] =current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
71
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: lowerCAmelCase__ :Any = None lowerCAmelCase__ :List[str] = logging.get_logger(__name__) lowerCAmelCase__ :Any = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ :Any = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ :Dict = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } lowerCAmelCase__ :Union[str, Any] = '''▁''' # Segments (not really needed) lowerCAmelCase__ :int = 0 lowerCAmelCase__ :Optional[Any] = 1 lowerCAmelCase__ :List[str] = 2 lowerCAmelCase__ :Optional[int] = 3 lowerCAmelCase__ :Optional[int] = 4 class __a ( UpperCAmelCase ): _a : int = VOCAB_FILES_NAMES _a : str = PRETRAINED_VOCAB_FILES_MAP _a : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a : Optional[int] = 'left' _a : Optional[Any] = XLNetTokenizer def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<sep>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE="<cls>" , _SCREAMING_SNAKE_CASE="<mask>" , _SCREAMING_SNAKE_CASE=["<eop>", "<eod>"] , **_SCREAMING_SNAKE_CASE , ) -> str: """simple docstring""" _UpperCAmelCase = AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else mask_token super().__init__( vocab_file=_SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE , remove_space=_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = 3 _UpperCAmelCase = do_lower_case _UpperCAmelCase = remove_space _UpperCAmelCase = keep_accents _UpperCAmelCase = vocab_file _UpperCAmelCase = False if not self.vocab_file else True def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[int]: """simple docstring""" _UpperCAmelCase = [self.sep_token_id] _UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[int]: """simple docstring""" _UpperCAmelCase = [self.sep_token_id] _UpperCAmelCase = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCAmelCase = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
350
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__ :Any = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :Optional[Any] = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :str = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :List[str] = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :str = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :List[Any] = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys lowerCAmelCase__ :Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
185
0
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer __lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) __lowerCAmelCase : Tuple = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} __lowerCAmelCase : Any = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } __lowerCAmelCase : Union[str, Any] = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } __lowerCAmelCase : Any = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } __lowerCAmelCase : Union[str, Any] = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } __lowerCAmelCase : Tuple = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } __lowerCAmelCase : Union[str, Any] = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } __lowerCAmelCase : Union[str, Any] = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } __lowerCAmelCase : Tuple = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } __lowerCAmelCase : str = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : Dict = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Optional[int] = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : List[str] = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE_ : Dict = DPRContextEncoderTokenizer class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : str = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Union[str, Any] = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Dict = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE_ : Optional[int] = DPRQuestionEncoderTokenizer __lowerCAmelCase : List[str] = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) __lowerCAmelCase : Optional[int] = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) __lowerCAmelCase : Optional[Any] = r'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(_UpperCamelCase ) class snake_case__ : """simple docstring""" def __call__( self : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Union[bool, str] = False , __lowerCamelCase : Union[bool, str] = False , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , __lowerCamelCase : Optional[bool] = None , **__lowerCamelCase : Optional[int] , ) -> BatchEncoding: if titles is None and texts is None: return super().__call__( __lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=__lowerCamelCase , return_tensors=__lowerCamelCase , return_attention_mask=__lowerCamelCase , **__lowerCamelCase , ) elif titles is None or texts is None: a = titles if texts is None else texts return super().__call__( __lowerCamelCase , __lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=__lowerCamelCase , return_tensors=__lowerCamelCase , return_attention_mask=__lowerCamelCase , **__lowerCamelCase , ) a = titles if not isinstance(__lowerCamelCase , __lowerCamelCase ) else [titles] a = texts if not isinstance(__lowerCamelCase , __lowerCamelCase ) else [texts] a = len(__lowerCamelCase ) a = questions if not isinstance(__lowerCamelCase , __lowerCamelCase ) else [questions] * n_passages assert len(__lowerCamelCase ) == len( __lowerCamelCase ), f"""There should be as many titles than texts but got {len(__lowerCamelCase )} titles and {len(__lowerCamelCase )} texts.""" a = super().__call__(__lowerCamelCase , __lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase )["input_ids"] a = super().__call__(__lowerCamelCase , add_special_tokens=__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase )["input_ids"] a = { "input_ids": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(__lowerCamelCase , __lowerCamelCase ) ] } if return_attention_mask is not False: a = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) a = attention_mask return self.pad(__lowerCamelCase , padding=__lowerCamelCase , max_length=__lowerCamelCase , return_tensors=__lowerCamelCase ) def __UpperCAmelCase ( self : Dict , __lowerCamelCase : BatchEncoding , __lowerCamelCase : DPRReaderOutput , __lowerCamelCase : int = 16 , __lowerCamelCase : int = 64 , __lowerCamelCase : int = 4 , ) -> List[DPRSpanPrediction]: a = reader_input["input_ids"] a , a , a = reader_output[:3] a = len(__lowerCamelCase ) a = sorted(range(__lowerCamelCase ) , reverse=__lowerCamelCase , key=relevance_logits.__getitem__ ) a = [] for doc_id in sorted_docs: a = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence a = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: a = sequence_ids.index(self.pad_token_id ) else: a = len(__lowerCamelCase ) a = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=__lowerCamelCase , top_spans=__lowerCamelCase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=__lowerCamelCase , start_index=__lowerCamelCase , end_index=__lowerCamelCase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(__lowerCamelCase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def __UpperCAmelCase ( self : List[str] , __lowerCamelCase : List[int] , __lowerCamelCase : List[int] , __lowerCamelCase : int , __lowerCamelCase : int , ) -> List[DPRSpanPrediction]: a = [] for start_index, start_score in enumerate(__lowerCamelCase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) a = sorted(__lowerCamelCase , key=lambda __lowerCamelCase : x[1] , reverse=__lowerCamelCase ) a = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f"""Wrong span indices: [{start_index}:{end_index}]""" a = end_index - start_index + 1 assert length <= max_answer_length, f"""Span is too long: {length} > {max_answer_length}""" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(__lowerCamelCase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(_UpperCamelCase ) class snake_case__ (_UpperCamelCase , _UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : List[str] = READER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Union[str, Any] = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : int = READER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE_ : int = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE_ : Optional[Any] = DPRReaderTokenizer
107
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Dict: # Initialise PyTorch model lowerCamelCase : Any = TaConfig.from_json_file(_SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) lowerCamelCase : str = TaForConditionalGeneration(_SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_ta(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) SCREAMING_SNAKE_CASE__ : str = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
48
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() a : Union[str, Any] = logging.get_logger(__name__) def lowercase__(A , A=False , A=False ) ->Optional[int]: """simple docstring""" lowercase__ : List[Any]= "backbone." if is_semantic else "" lowercase__ : Tuple= [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', "beit.embeddings.cls_token"), (f'''{prefix}patch_embed.proj.weight''', "beit.embeddings.patch_embeddings.projection.weight"), (f'''{prefix}patch_embed.proj.bias''', "beit.embeddings.patch_embeddings.projection.bias"), (f'''{prefix}pos_embed''', "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def lowercase__(A , A , A=False , A=False ) ->Union[str, Any]: """simple docstring""" for i in range(config.num_hidden_layers ): lowercase__ : Union[str, Any]= "backbone." if is_semantic else "" # queries, keys and values lowercase__ : Dict= state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) lowercase__ : Dict= state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) lowercase__ : List[Any]= state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) lowercase__ : int= in_proj_weight[ : config.hidden_size, : ] lowercase__ : Tuple= q_bias lowercase__ : Dict= in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase__ : List[Any]= in_proj_weight[ -config.hidden_size :, : ] lowercase__ : List[str]= v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained lowercase__ : str= state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) lowercase__ : Union[str, Any]= state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) lowercase__ : List[str]= gamma_a lowercase__ : Any= gamma_a def lowercase__(A , A , A ) ->Tuple: """simple docstring""" lowercase__ : int= dct.pop(A ) lowercase__ : Optional[Any]= val def lowercase__() ->Tuple: """simple docstring""" lowercase__ : Dict= "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ : List[Any]= Image.open(requests.get(A , stream=A ).raw ) return im @torch.no_grad() def lowercase__(A , A , A=False ) ->int: """simple docstring""" lowercase__ : List[str]= False if "rvlcdip" in checkpoint_url else True lowercase__ : Union[str, Any]= BeitConfig(use_absolute_position_embeddings=A , use_mask_token=A ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: lowercase__ : int= 1_024 lowercase__ : List[str]= 4_096 lowercase__ : List[str]= 24 lowercase__ : Optional[int]= 16 # labels if "rvlcdip" in checkpoint_url: lowercase__ : Optional[Any]= 16 lowercase__ : Dict= "huggingface/label-files" lowercase__ : Union[str, Any]= "rvlcdip-id2label.json" lowercase__ : str= json.load(open(hf_hub_download(A , A , repo_type="dataset" ) , "r" ) ) lowercase__ : Optional[int]= {int(A ): v for k, v in idalabel.items()} lowercase__ : Tuple= idalabel lowercase__ : Union[str, Any]= {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys lowercase__ : str= torch.hub.load_state_dict_from_url(A , map_location="cpu" )["model"] lowercase__ : str= create_rename_keys(A , has_lm_head=A ) for src, dest in rename_keys: rename_key(A , A , A ) read_in_q_k_v(A , A , has_lm_head=A ) # load HuggingFace model lowercase__ : List[str]= BeitForMaskedImageModeling(A ) if has_lm_head else BeitForImageClassification(A ) model.eval() model.load_state_dict(A ) # Check outputs on an image lowercase__ : Optional[Any]= BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=A ) lowercase__ : int= prepare_img() lowercase__ : Tuple= image_processor(images=A , return_tensors="pt" ) lowercase__ : Dict= encoding["pixel_values"] lowercase__ : Union[str, Any]= model(A ) lowercase__ : Dict= outputs.logits # verify logits lowercase__ : Union[str, Any]= [1, 16] if "rvlcdip" in checkpoint_url else [1, 196, 8_192] assert logits.shape == torch.Size(A ), "Shape of logits not as expected" Path(A ).mkdir(exist_ok=A ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(A ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(A ) if push_to_hub: if has_lm_head: lowercase__ : List[str]= "dit-base" if "base" in checkpoint_url else "dit-large" else: lowercase__ : Optional[Any]= "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(A , A ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=A , ) model.push_to_hub( repo_path_or_name=Path(A , A ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=A , ) if __name__ == "__main__": a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) a : Optional[int] = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
150
"""simple docstring""" import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py a : Any = """src/diffusers""" # Matches is_xxx_available() a : Optional[Any] = re.compile(r"""is\_([a-z_]*)_available\(\)""") # Matches from xxx import bla a : Dict = re.compile(r"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") a : Dict = """ {0} = None """ a : List[Any] = """ class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, {1}) @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, {1}) """ a : Any = """ def {0}(*args, **kwargs): requires_backends({0}, {1}) """ def lowercase__(A ) ->List[str]: """simple docstring""" lowercase__ : Optional[int]= _re_backend.findall(A ) if len(A ) == 0: return None return "_and_".join(A ) def lowercase__() ->int: """simple docstring""" with open(os.path.join(A , "__init__.py" ) , "r" , encoding="utf-8" , newline="\n" ) as f: lowercase__ : Union[str, Any]= f.readlines() # Get to the point we do the actual imports for type checking lowercase__ : List[Any]= 0 lowercase__ : List[Any]= {} # Go through the end of the file while line_index < len(A ): # If the line contains is_backend_available, we grab all objects associated with the `else` block lowercase__ : Optional[Any]= find_backend(lines[line_index] ) if backend is not None: while not lines[line_index].startswith("else:" ): line_index += 1 line_index += 1 lowercase__ : str= [] # Until we unindent, add backend objects to the list while line_index < len(A ) and len(lines[line_index] ) > 1: lowercase__ : List[str]= lines[line_index] lowercase__ : Any= _re_single_line_import.search(A ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 if len(A ) > 0: lowercase__ : str= objects else: line_index += 1 return backend_specific_objects def lowercase__(A , A ) ->List[Any]: """simple docstring""" if name.isupper(): return DUMMY_CONSTANT.format(A ) elif name.islower(): return DUMMY_FUNCTION.format(A , A ) else: return DUMMY_CLASS.format(A , A ) def lowercase__(A=None ) ->Optional[Any]: """simple docstring""" if backend_specific_objects is None: lowercase__ : int= read_init() # For special correspondence backend to module name as used in the function requires_modulename lowercase__ : Dict= {} for backend, objects in backend_specific_objects.items(): lowercase__ : str= "[" + ", ".join(f'''"{b}"''' for b in backend.split("_and_" ) ) + "]" lowercase__ : List[str]= "# This file is autogenerated by the command `make fix-copies`, do not edit.\n" dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(A , A ) for o in objects] ) lowercase__ : Optional[Any]= dummy_file return dummy_files def lowercase__(A=False ) ->List[Any]: """simple docstring""" lowercase__ : Tuple= create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py lowercase__ : int= {"torch": "pt"} # Locate actual dummy modules and read their content. lowercase__ : Dict= os.path.join(A , "utils" ) lowercase__ : str= { backend: os.path.join(A , f'''dummy_{short_names.get(A , A )}_objects.py''' ) for backend in dummy_files.keys() } lowercase__ : Union[str, Any]= {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(A ): with open(A , "r" , encoding="utf-8" , newline="\n" ) as f: lowercase__ : List[Any]= f.read() else: lowercase__ : Tuple= "" for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( f'''Updating diffusers.utils.dummy_{short_names.get(A , A )}_objects.py as the main ''' "__init__ has new objects." ) with open(dummy_file_paths[backend] , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(dummy_files[backend] ) else: raise ValueError( "The main __init__ has objects that are not present in " f'''diffusers.utils.dummy_{short_names.get(A , A )}_objects.py. Run `make fix-copies` ''' "to fix this." ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") a : Tuple = parser.parse_args() check_dummies(args.fix_and_overwrite)
150
1
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
126
"""simple docstring""" import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase = logging.get_logger(__name__) def lowerCAmelCase_ ( snake_case_ : List[str] ) ->List[str]: print('Loading config file...' ) def flatten_yaml_as_dict(snake_case_ : str , snake_case_ : Optional[int]="" , snake_case_ : Optional[Any]="." ): lowerCamelCase__ : Union[str, Any] =[] for k, v in d.items(): lowerCamelCase__ : Any =parent_key + sep + k if parent_key else k if isinstance(snake_case_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(snake_case_ , snake_case_ , sep=snake_case_ ).items() ) else: items.append((new_key, v) ) return dict(snake_case_ ) lowerCamelCase__ : str =argparse.Namespace() with open(snake_case_ , 'r' ) as yaml_file: try: lowerCamelCase__ : Union[str, Any] =yaml.load(snake_case_ , Loader=yaml.FullLoader ) lowerCamelCase__ : List[Any] =flatten_yaml_as_dict(snake_case_ ) for k, v in flat_cfg.items(): setattr(snake_case_ , snake_case_ , snake_case_ ) except yaml.YAMLError as exc: logger.error('Error while loading config file: {}. Error message: {}'.format(snake_case_ , str(snake_case_ ) ) ) return config def lowerCAmelCase_ ( snake_case_ : Dict , snake_case_ : Optional[Any] ) ->List[Any]: lowerCamelCase__ : Tuple =MobileViTVaConfig() lowerCamelCase__ : List[Any] =False # dataset if task_name.startswith('imagenet1k_' ): lowerCamelCase__ : List[str] =1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: lowerCamelCase__ : Optional[int] =3_8_4 else: lowerCamelCase__ : Dict =2_5_6 lowerCamelCase__ : Dict ='imagenet-1k-id2label.json' elif task_name.startswith('imagenet21k_to_1k_' ): lowerCamelCase__ : Tuple =2_1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: lowerCamelCase__ : Tuple =3_8_4 else: lowerCamelCase__ : List[str] =2_5_6 lowerCamelCase__ : Any ='imagenet-22k-id2label.json' elif task_name.startswith('ade20k_' ): lowerCamelCase__ : int =1_5_1 lowerCamelCase__ : int =5_1_2 lowerCamelCase__ : Dict ='ade20k-id2label.json' lowerCamelCase__ : Union[str, Any] =True elif task_name.startswith('voc_' ): lowerCamelCase__ : Optional[Any] =2_1 lowerCamelCase__ : Dict =5_1_2 lowerCamelCase__ : Union[str, Any] ='pascal-voc-id2label.json' lowerCamelCase__ : int =True # orig_config lowerCamelCase__ : Union[str, Any] =load_orig_config_file(snake_case_ ) assert getattr(snake_case_ , 'model.classification.name' , -1 ) == "mobilevit_v2", "Invalid model" lowerCamelCase__ : List[str] =getattr(snake_case_ , 'model.classification.mitv2.width_multiplier' , 1.0 ) assert ( getattr(snake_case_ , 'model.classification.mitv2.attn_norm_layer' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" lowerCamelCase__ : List[Any] =getattr(snake_case_ , 'model.classification.activation.name' , 'swish' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: lowerCamelCase__ : str =getattr(snake_case_ , 'model.segmentation.output_stride' , 1_6 ) if "_deeplabv3" in task_name: lowerCamelCase__ : Dict =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_rates' , [1_2, 2_4, 3_6] ) lowerCamelCase__ : str =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_out_channels' , 5_1_2 ) lowerCamelCase__ : Dict =getattr(snake_case_ , 'model.segmentation.deeplabv3.aspp_dropout' , 0.1 ) # id2label lowerCamelCase__ : Union[str, Any] ='huggingface/label-files' lowerCamelCase__ : List[Any] =json.load(open(hf_hub_download(snake_case_ , snake_case_ , repo_type='dataset' ) , 'r' ) ) lowerCamelCase__ : List[str] ={int(snake_case_ ): v for k, v in idalabel.items()} lowerCamelCase__ : Optional[int] =idalabel lowerCamelCase__ : List[str] ={v: k for k, v in idalabel.items()} return config def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : List[str] , snake_case_ : Tuple ) ->Tuple: lowerCamelCase__ : List[Any] =dct.pop(snake_case_ ) lowerCamelCase__ : str =val def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Any=False ) ->str: if base_model: lowerCamelCase__ : Tuple ='' else: lowerCamelCase__ : Optional[Any] ='mobilevitv2.' lowerCamelCase__ : Tuple =[] for k in state_dict.keys(): if k[:8] == "encoder.": lowerCamelCase__ : Tuple =k[8:] else: lowerCamelCase__ : Tuple =k if ".block." in k: lowerCamelCase__ : Optional[Any] =k_new.replace('.block.' , '.' ) if ".conv." in k: lowerCamelCase__ : Union[str, Any] =k_new.replace('.conv.' , '.convolution.' ) if ".norm." in k: lowerCamelCase__ : Dict =k_new.replace('.norm.' , '.normalization.' ) if "conv_1." in k: lowerCamelCase__ : Dict =k_new.replace('conv_1.' , f"""{model_prefix}conv_stem.""" ) for i in [1, 2]: if f"""layer_{i}.""" in k: lowerCamelCase__ : Union[str, Any] =k_new.replace(f"""layer_{i}.""" , f"""{model_prefix}encoder.layer.{i-1}.layer.""" ) if ".exp_1x1." in k: lowerCamelCase__ : Tuple =k_new.replace('.exp_1x1.' , '.expand_1x1.' ) if ".red_1x1." in k: lowerCamelCase__ : Union[str, Any] =k_new.replace('.red_1x1.' , '.reduce_1x1.' ) for i in [3, 4, 5]: if f"""layer_{i}.0.""" in k: lowerCamelCase__ : Tuple =k_new.replace(f"""layer_{i}.0.""" , f"""{model_prefix}encoder.layer.{i-1}.downsampling_layer.""" ) if f"""layer_{i}.1.local_rep.0.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.local_rep.0.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_kxk.""" ) if f"""layer_{i}.1.local_rep.1.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.local_rep.1.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_1x1.""" ) for i in [3, 4, 5]: if i == 3: lowerCamelCase__ : Optional[int] =[0, 1] elif i == 4: lowerCamelCase__ : int =[0, 1, 2, 3] elif i == 5: lowerCamelCase__ : List[str] =[0, 1, 2] for j in j_in: if f"""layer_{i}.1.global_rep.{j}.""" in k: lowerCamelCase__ : Any =k_new.replace( f"""layer_{i}.1.global_rep.{j}.""" , f"""{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.""" ) if f"""layer_{i}.1.global_rep.{j+1}.""" in k: lowerCamelCase__ : Optional[Any] =k_new.replace( f"""layer_{i}.1.global_rep.{j+1}.""" , f"""{model_prefix}encoder.layer.{i-1}.layernorm.""" ) if f"""layer_{i}.1.conv_proj.""" in k: lowerCamelCase__ : Any =k_new.replace(f"""layer_{i}.1.conv_proj.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_projection.""" ) if "pre_norm_attn.0." in k: lowerCamelCase__ : int =k_new.replace('pre_norm_attn.0.' , 'layernorm_before.' ) if "pre_norm_attn.1." in k: lowerCamelCase__ : Optional[int] =k_new.replace('pre_norm_attn.1.' , 'attention.' ) if "pre_norm_ffn.0." in k: lowerCamelCase__ : List[str] =k_new.replace('pre_norm_ffn.0.' , 'layernorm_after.' ) if "pre_norm_ffn.1." in k: lowerCamelCase__ : Any =k_new.replace('pre_norm_ffn.1.' , 'ffn.conv1.' ) if "pre_norm_ffn.3." in k: lowerCamelCase__ : Tuple =k_new.replace('pre_norm_ffn.3.' , 'ffn.conv2.' ) if "classifier.1." in k: lowerCamelCase__ : List[Any] =k_new.replace('classifier.1.' , 'classifier.' ) if "seg_head." in k: lowerCamelCase__ : int =k_new.replace('seg_head.' , 'segmentation_head.' ) if ".aspp_layer." in k: lowerCamelCase__ : List[str] =k_new.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in k: lowerCamelCase__ : Optional[int] =k_new.replace('.aspp_pool.' , '.' ) rename_keys.append((k, k_new) ) return rename_keys def lowerCAmelCase_ ( snake_case_ : Tuple ) ->str: lowerCamelCase__ : str =[] for k in state_dict.keys(): if k.startswith('seg_head.aux_head.' ): keys_to_ignore.append(snake_case_ ) for k in keys_to_ignore: state_dict.pop(snake_case_ , snake_case_ ) def lowerCAmelCase_ ( ) ->Optional[int]: lowerCamelCase__ : List[str] ='http://images.cocodataset.org/val2017/000000039769.jpg' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" lowerCamelCase__ : Any =Image.open(requests.get(snake_case_ , stream=snake_case_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Dict , snake_case_ : Any , snake_case_ : List[str] ) ->Optional[int]: lowerCamelCase__ : int =get_mobilevitva_config(snake_case_ , snake_case_ ) # load original state_dict lowerCamelCase__ : List[str] =torch.load(snake_case_ , map_location='cpu' ) # load huggingface model if task_name.startswith('ade20k_' ) or task_name.startswith('voc_' ): lowerCamelCase__ : List[str] =MobileViTVaForSemanticSegmentation(snake_case_ ).eval() lowerCamelCase__ : Optional[Any] =False else: lowerCamelCase__ : Union[str, Any] =MobileViTVaForImageClassification(snake_case_ ).eval() lowerCamelCase__ : List[Any] =False # remove and rename some keys of load the original model lowerCamelCase__ : Union[str, Any] =checkpoint remove_unused_keys(snake_case_ ) lowerCamelCase__ : List[str] =create_rename_keys(snake_case_ , base_model=snake_case_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(snake_case_ , snake_case_ , snake_case_ ) # load modified state_dict model.load_state_dict(snake_case_ ) # Check outputs on an image, prepared by MobileViTImageProcessor lowerCamelCase__ : str =MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 3_2 ) lowerCamelCase__ : int =image_processor(images=prepare_img() , return_tensors='pt' ) lowerCamelCase__ : Optional[int] =model(**snake_case_ ) # verify classification model if task_name.startswith('imagenet' ): lowerCamelCase__ : Tuple =outputs.logits lowerCamelCase__ : Optional[int] =logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('imagenet1k_256' ) and config.width_multiplier == 1.0: # expected_logits for base variant lowerCamelCase__ : Tuple =torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ) assert torch.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) Path(snake_case_ ).mkdir(exist_ok=snake_case_ ) print(f"""Saving model {task_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(snake_case_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(snake_case_ ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you'd like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCAmelCase = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
126
1
import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def lowerCAmelCase__ ( _a : Union[dict, list, tuple, torch.Tensor] ): snake_case_ : List[str] = [] if isinstance(_a , _a ): for v in tree.values(): shapes.extend(_fetch_dims(_a ) ) elif isinstance(_a , (list, tuple) ): for t in tree: shapes.extend(_fetch_dims(_a ) ) elif isinstance(_a , torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError("Not supported" ) return shapes @torch.jit.ignore def lowerCAmelCase__ ( _a : int , _a : Tuple[int, ...] ): snake_case_ : List[str] = [] for d in reversed(_a ): idx.append(flat_idx % d ) snake_case_ : List[Any] = flat_idx // d return tuple(reversed(_a ) ) @torch.jit.ignore def lowerCAmelCase__ ( _a : Sequence[int] , _a : Sequence[int] , _a : Sequence[int] , _a : Optional[Sequence[bool]] = None , _a : Optional[Sequence[bool]] = None , ): # start_edges and end_edges both indicate whether, starting from any given # dimension, the start/end index is at the top/bottom edge of the # corresponding tensor, modeled as a tree def reduce_edge_list(_a : List[bool] ) -> None: snake_case_ : Tuple = True for i in range(len(_a ) ): snake_case_ : Any = -1 * (i + 1) l[reversed_idx] &= tally snake_case_ : Tuple = l[reversed_idx] if start_edges is None: snake_case_ : List[Any] = [s == 0 for s in start] reduce_edge_list(_a ) if end_edges is None: snake_case_ : str = [e == (d - 1) for e, d in zip(_a , _a )] reduce_edge_list(_a ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(_a ) == 0: return [()] elif len(_a ) == 1: return [(slice(start[0] , end[0] + 1 ),)] snake_case_ : List[Tuple[slice, ...]] = [] snake_case_ : List[slice] = [] # Dimensions common to start and end can be selected directly for s, e in zip(_a , _a ): if s == e: path_list.append(slice(_a , s + 1 ) ) else: break snake_case_ : Tuple[slice, ...] = tuple(_a ) snake_case_ : Tuple = len(_a ) # start == end, and we're done if divergence_idx == len(_a ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None snake_case_ : int = start[divergence_idx] return tuple( path + (slice(_a , sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] , [d - 1 for d in dims[divergence_idx + 1 :]] , dims[divergence_idx + 1 :] , start_edges=start_edges[divergence_idx + 1 :] , end_edges=[True for _ in end_edges[divergence_idx + 1 :]] , ) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None snake_case_ : Any = end[divergence_idx] return tuple( path + (slice(_a , edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] , end[divergence_idx + 1 :] , dims[divergence_idx + 1 :] , start_edges=[True for _ in start_edges[divergence_idx + 1 :]] , end_edges=end_edges[divergence_idx + 1 :] , ) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) snake_case_ : str = end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def lowerCAmelCase__ ( _a : torch.Tensor , _a : int , _a : int , _a : int ): snake_case_ : Dict = t.shape[:no_batch_dims] snake_case_ : List[Any] = list(_flat_idx_to_idx(_a , _a ) ) # _get_minimal_slice_set is inclusive snake_case_ : Tuple = list(_flat_idx_to_idx(flat_end - 1 , _a ) ) # Get an ordered list of slices to perform snake_case_ : int = _get_minimal_slice_set( _a , _a , _a , ) snake_case_ : List[Any] = [t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def lowerCAmelCase__ ( _a : Callable , _a : Dict[str, Any] , _a : int , _a : int , _a : bool = False , _a : Any = None , _a : bool = False , ): if not (len(_a ) > 0): raise ValueError("Must provide at least one input" ) snake_case_ : Tuple = [shape[:no_batch_dims] for shape in _fetch_dims(_a )] snake_case_ : Dict = tuple([max(_a ) for s in zip(*_a )] ) def _prep_inputs(_a : torch.Tensor ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: snake_case_ : Optional[int] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) snake_case_ : int = t.reshape(-1 , *t.shape[no_batch_dims:] ) else: snake_case_ : Tuple = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t snake_case_ : Dict[str, Any] = tensor_tree_map(_prep_inputs , _a ) snake_case_ : Any = None if _out is not None: snake_case_ : int = tensor_tree_map(lambda _a : t.view([-1] + list(t.shape[no_batch_dims:] ) ) , _out ) snake_case_ : Any = 1 for d in orig_batch_dims: flat_batch_dim *= d snake_case_ : str = flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(_a : torch.Tensor ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t snake_case_ : str = 0 snake_case_ : Dict = prepped_outputs for _ in range(_a ): # Chunk the input if not low_mem: snake_case_ : int = _select_chunk else: snake_case_ : List[str] = partial( _chunk_slice , flat_start=_a , flat_end=min(_a , i + chunk_size ) , no_batch_dims=len(_a ) , ) snake_case_ : Dict[str, Any] = tensor_tree_map(_a , _a ) # Run the layer on the chunk snake_case_ : Optional[int] = layer(**_a ) # Allocate space for the output if out is None: snake_case_ : Union[str, Any] = tensor_tree_map(lambda _a : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) , _a ) # Put the chunk in its pre-allocated space if isinstance(_a , _a ): def assign(_a : dict , _a : dict ) -> None: for k, v in da.items(): if isinstance(_a , _a ): assign(_a , da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: snake_case_ : Union[str, Any] = da[k] assign(_a , _a ) elif isinstance(_a , _a ): for xa, xa in zip(_a , _a ): if _add_into_out: xa[i : i + chunk_size] += xa else: snake_case_ : List[str] = xa elif isinstance(_a , torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: snake_case_ : Tuple = output_chunk else: raise ValueError("Not supported" ) i += chunk_size snake_case_ : Optional[Any] = tensor_tree_map(lambda _a : t.view(orig_batch_dims + t.shape[1:] ) , _a ) return out class UpperCAmelCase_ : '''simple docstring''' def __init__( self , _SCREAMING_SNAKE_CASE = 512 , ) -> List[Any]: snake_case_ : int = max_chunk_size snake_case_ : Optional[int] = None snake_case_ : Optional[tuple] = None def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: logging.info("Tuning chunk size..." ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size snake_case_ : List[int] = [2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] snake_case_ : Optional[Any] = [c for c in candidates if c > min_chunk_size] snake_case_ : Any = [min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(_SCREAMING_SNAKE_CASE ) -> bool: try: with torch.no_grad(): fn(*_SCREAMING_SNAKE_CASE , chunk_size=_SCREAMING_SNAKE_CASE ) return True except RuntimeError: return False snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) - 1 while i > min_viable_chunk_size_index: snake_case_ : Union[str, Any] = test_chunk_size(candidates[i] ) if not viable: snake_case_ : Dict = (min_viable_chunk_size_index + i) // 2 else: snake_case_ : Optional[Any] = i snake_case_ : int = (i + len(_SCREAMING_SNAKE_CASE ) - 1) // 2 return candidates[min_viable_chunk_size_index] def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> bool: snake_case_ : List[str] = True for aa, aa in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): assert type(_SCREAMING_SNAKE_CASE ) == type(_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , (list, tuple) ): consistent &= self._compare_arg_caches(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): snake_case_ : List[Any] = [v for _, v in sorted(aa.items() , key=lambda _SCREAMING_SNAKE_CASE : x[0] )] snake_case_ : List[Any] = [v for _, v in sorted(aa.items() , key=lambda _SCREAMING_SNAKE_CASE : x[0] )] consistent &= self._compare_arg_caches(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: consistent &= aa == aa return consistent def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> int: snake_case_ : List[Any] = True snake_case_ : tuple = tree_map(lambda _SCREAMING_SNAKE_CASE : a.shape if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ) else a , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(_SCREAMING_SNAKE_CASE ) snake_case_ : Union[str, Any] = self._compare_arg_caches(self.cached_arg_data , _SCREAMING_SNAKE_CASE ) else: # Otherwise, we can reuse the precomputed value snake_case_ : List[Any] = False if not consistent: snake_case_ : List[Any] = self._determine_favorable_chunk_size( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) snake_case_ : List[str] = arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
358
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> int: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["bs"] , model_result["ss"] ): snake_case_ : List[Any] = model_result["result"][batch_size][sequence_length] self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self ) -> Tuple: snake_case_ : Dict = "sshleifer/tiny-gpt2" snake_case_ : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=_SCREAMING_SNAKE_CASE , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : Optional[Any] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> int: snake_case_ : List[Any] = "sgugger/tiny-distilbert-classification" snake_case_ : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , only_pretrain_model=_SCREAMING_SNAKE_CASE , ) snake_case_ : int = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> Tuple: snake_case_ : List[str] = "sshleifer/tiny-gpt2" snake_case_ : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : Optional[Any] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> int: snake_case_ : Union[str, Any] = "sshleifer/tiny-gpt2" snake_case_ : List[str] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) snake_case_ : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=_SCREAMING_SNAKE_CASE , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : List[str] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE , [config] ) snake_case_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> List[str]: snake_case_ : str = "sshleifer/tiny-gpt2" snake_case_ : int = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) snake_case_ : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : List[Any] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE , [config] ) snake_case_ : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> str: snake_case_ : List[str] = "sshleifer/tiny-gpt2" snake_case_ : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : List[str] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowerCAmelCase ( self ) -> Dict: snake_case_ : str = "sshleifer/tiny-gpt2" snake_case_ : str = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) snake_case_ : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : Optional[Any] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE , [config] ) snake_case_ : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowerCAmelCase ( self ) -> List[str]: snake_case_ : List[str] = "patrickvonplaten/t5-tiny-random" snake_case_ : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : List[str] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE , configs=[config] ) snake_case_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices("GPU" ) ) == 0 , "Cannot do xla on CPU." ) def _lowerCAmelCase ( self ) -> Dict: snake_case_ : int = "sshleifer/tiny-gpt2" snake_case_ : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=_SCREAMING_SNAKE_CASE , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , use_xla=_SCREAMING_SNAKE_CASE , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : List[str] = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowerCAmelCase ( self ) -> Tuple: snake_case_ : Union[str, Any] = "sshleifer/tiny-gpt2" with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=_SCREAMING_SNAKE_CASE , save_to_csv=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(_SCREAMING_SNAKE_CASE , "inf_time.csv" ) , inference_memory_csv_file=os.path.join(_SCREAMING_SNAKE_CASE , "inf_mem.csv" ) , env_info_csv_file=os.path.join(_SCREAMING_SNAKE_CASE , "env.csv" ) , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : Dict = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) benchmark.run() self.assertTrue(Path(os.path.join(_SCREAMING_SNAKE_CASE , "inf_time.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(_SCREAMING_SNAKE_CASE , "inf_mem.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(_SCREAMING_SNAKE_CASE , "env.csv" ) ).exists() ) def _lowerCAmelCase ( self ) -> List[str]: snake_case_ : int = "sshleifer/tiny-gpt2" def _check_summary_is_not_empty(_SCREAMING_SNAKE_CASE ): self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "sequential" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "cumulative" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "current" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "total" ) ) with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=_SCREAMING_SNAKE_CASE , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(_SCREAMING_SNAKE_CASE , "log.txt" ) , log_print=_SCREAMING_SNAKE_CASE , trace_memory_line_by_line=_SCREAMING_SNAKE_CASE , eager_mode=_SCREAMING_SNAKE_CASE , multi_process=_SCREAMING_SNAKE_CASE , ) snake_case_ : Tuple = TensorFlowBenchmark(_SCREAMING_SNAKE_CASE ) snake_case_ : int = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(_SCREAMING_SNAKE_CASE , "log.txt" ) ).exists() )
36
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a =logging.get_logger(__name__) a ={ """caidas/swin2sr-classicalsr-x2-64""": ( """https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json""" ), } class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : int = '''swin2sr''' _UpperCAmelCase : Dict = { '''hidden_size''': '''embed_dim''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Tuple ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=6_4 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=1 ,SCREAMING_SNAKE_CASE__ : Optional[Any]=3 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=1_8_0 ,SCREAMING_SNAKE_CASE__ : Tuple=[6, 6, 6, 6, 6, 6] ,SCREAMING_SNAKE_CASE__ : Tuple=[6, 6, 6, 6, 6, 6] ,SCREAMING_SNAKE_CASE__ : Optional[Any]=8 ,SCREAMING_SNAKE_CASE__ : Any=2.0 ,SCREAMING_SNAKE_CASE__ : str=True ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.0 ,SCREAMING_SNAKE_CASE__ : Optional[Any]=0.0 ,SCREAMING_SNAKE_CASE__ : Tuple=0.1 ,SCREAMING_SNAKE_CASE__ : Any="gelu" ,SCREAMING_SNAKE_CASE__ : Dict=False ,SCREAMING_SNAKE_CASE__ : Dict=0.02 ,SCREAMING_SNAKE_CASE__ : Any=1E-5 ,SCREAMING_SNAKE_CASE__ : Dict=2 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=1.0 ,SCREAMING_SNAKE_CASE__ : str="1conv" ,SCREAMING_SNAKE_CASE__ : List[str]="pixelshuffle" ,**SCREAMING_SNAKE_CASE__ : Any ,): super().__init__(**SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = image_size __lowerCamelCase : int = patch_size __lowerCamelCase : Any = num_channels __lowerCamelCase : Optional[int] = embed_dim __lowerCamelCase : int = depths __lowerCamelCase : Optional[int] = len(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Any = num_heads __lowerCamelCase : Tuple = window_size __lowerCamelCase : Optional[int] = mlp_ratio __lowerCamelCase : List[Any] = qkv_bias __lowerCamelCase : Tuple = hidden_dropout_prob __lowerCamelCase : str = attention_probs_dropout_prob __lowerCamelCase : Union[str, Any] = drop_path_rate __lowerCamelCase : Tuple = hidden_act __lowerCamelCase : Union[str, Any] = use_absolute_embeddings __lowerCamelCase : Optional[Any] = layer_norm_eps __lowerCamelCase : Dict = initializer_range __lowerCamelCase : Any = upscale __lowerCamelCase : Any = img_range __lowerCamelCase : List[str] = resi_connection __lowerCamelCase : Optional[Any] = upsampler
73
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> int: __lowerCamelCase : Optional[int] = 0 __lowerCamelCase : Dict = len(lowerCamelCase__ ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCamelCase : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCamelCase__ ): return None __lowerCamelCase : Tuple = sorted_collection[point] if current_item == item: return point else: if point < left: __lowerCamelCase : List[Any] = left __lowerCamelCase : Tuple = point elif point > right: __lowerCamelCase : Dict = right __lowerCamelCase : str = point else: if item < current_item: __lowerCamelCase : Dict = point - 1 else: __lowerCamelCase : Dict = point + 1 return None def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Any: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCamelCase : Optional[int] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCamelCase__ ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) elif point > right: return interpolation_search_by_recursion(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , point - 1 ) else: return interpolation_search_by_recursion( lowerCamelCase__ , lowerCamelCase__ , point + 1 , lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Optional[Any]: if collection != sorted(lowerCamelCase__ ): raise ValueError('Collection must be ascending sorted' ) return True if __name__ == "__main__": import sys a =0 if debug == 1: a =[10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("""Sequence must be ascending sorted to apply interpolation search""") a =67 a =interpolation_search(collection, target) if result is not None: print(F"""{target} found at positions: {result}""") else: print("""Not found""")
73
1
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class A ( UpperCAmelCase_ , UpperCAmelCase_ ): __UpperCAmelCase : Optional[Any] = 1 @register_to_config def __init__(self : List[str] , __UpperCAmelCase : Union[str, Any]=2_0_0_0 , __UpperCAmelCase : str=0.1 , __UpperCAmelCase : Optional[Any]=2_0 , __UpperCAmelCase : str=1E-3 ) -> Dict: """simple docstring""" UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None def lowercase_ (self : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any] = None ) -> Any: """simple docstring""" UpperCAmelCase__ = torch.linspace(1 , self.config.sampling_eps , _a , device=_a ) def lowercase_ (self : Dict , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any]=None ) -> str: """simple docstring""" if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler" ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score UpperCAmelCase__ = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) UpperCAmelCase__ = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) UpperCAmelCase__ = std.flatten() while len(std.shape ) < len(score.shape ): UpperCAmelCase__ = std.unsqueeze(-1 ) UpperCAmelCase__ = -score / std # compute UpperCAmelCase__ = -1.0 / len(self.timesteps ) UpperCAmelCase__ = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) UpperCAmelCase__ = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): UpperCAmelCase__ = beta_t.unsqueeze(-1 ) UpperCAmelCase__ = -0.5 * beta_t * x UpperCAmelCase__ = torch.sqrt(_a ) UpperCAmelCase__ = drift - diffusion**2 * score UpperCAmelCase__ = x + drift * dt # add noise UpperCAmelCase__ = randn_tensor(x.shape , layout=x.layout , generator=_a , device=x.device , dtype=x.dtype ) UpperCAmelCase__ = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__(self : Dict ) -> Tuple: """simple docstring""" return self.config.num_train_timesteps
369
def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
143
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __lowerCAmelCase ( _UpperCAmelCase, unittest.TestCase ): lowerCamelCase_ : str = BertJapaneseTokenizer lowerCamelCase_ : Dict = False lowerCamelCase_ : Any = True def lowerCamelCase (self ) -> Tuple: '''simple docstring''' super().setUp() snake_case_ : int = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def lowerCamelCase (self , __magic_name__ ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = 'こんにちは、世界。 \nこんばんは、世界。' snake_case_ : Any = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def lowerCamelCase (self , __magic_name__ ) -> Optional[int]: '''simple docstring''' snake_case_ : Any = self.get_input_output_texts(SCREAMING_SNAKE_CASE_ ) snake_case_ : List[Any] = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) snake_case_ : Dict = tokenizer.decode(SCREAMING_SNAKE_CASE_ , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) return text, ids def lowerCamelCase (self ) -> List[str]: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : int = self.tokenizer_class(self.vocab_file ) snake_case_ : Union[str, Any] = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : int = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) snake_case_ : List[str] = 'こんにちは、世界。\nこんばんは、世界。' snake_case_ : Tuple = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case_ : Tuple = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(SCREAMING_SNAKE_CASE_ , '''wb''' ) as handle: pickle.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , '''rb''' ) as handle: snake_case_ : List[Any] = pickle.load(SCREAMING_SNAKE_CASE_ ) snake_case_ : List[str] = tokenizer_new.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' try: snake_case_ : Dict = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def lowerCamelCase (self ) -> int: '''simple docstring''' try: snake_case_ : List[Any] = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : str = MecabTokenizer(do_lower_case=SCREAMING_SNAKE_CASE_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def lowerCamelCase (self ) -> str: '''simple docstring''' try: snake_case_ : Any = MecabTokenizer( do_lower_case=SCREAMING_SNAKE_CASE_ , normalize_text=SCREAMING_SNAKE_CASE_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[Any] = MecabTokenizer(normalize_text=SCREAMING_SNAKE_CASE_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Any = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) snake_case_ : Optional[int] = 'こんにちは、世界。\nこんばんは、世界。' snake_case_ : Optional[Any] = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case_ : Optional[int] = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(SCREAMING_SNAKE_CASE_ , '''wb''' ) as handle: pickle.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , '''rb''' ) as handle: snake_case_ : Dict = pickle.load(SCREAMING_SNAKE_CASE_ ) snake_case_ : Optional[int] = tokenizer_new.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) @require_sudachi def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : Any = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : List[Any] = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = SudachiTokenizer(do_lower_case=SCREAMING_SNAKE_CASE_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = SudachiTokenizer(normalize_text=SCREAMING_SNAKE_CASE_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : Any = SudachiTokenizer(trim_whitespace=SCREAMING_SNAKE_CASE_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) snake_case_ : List[Any] = 'こんにちは、世界。\nこんばんは、世界。' snake_case_ : List[Any] = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case_ : str = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(SCREAMING_SNAKE_CASE_ , '''wb''' ) as handle: pickle.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , '''rb''' ) as handle: snake_case_ : List[Any] = pickle.load(SCREAMING_SNAKE_CASE_ ) snake_case_ : Optional[int] = tokenizer_new.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) @require_jumanpp def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Dict = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : str = JumanppTokenizer(do_lower_case=SCREAMING_SNAKE_CASE_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : List[str] = JumanppTokenizer(normalize_text=SCREAMING_SNAKE_CASE_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Dict = JumanppTokenizer(trim_whitespace=SCREAMING_SNAKE_CASE_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : List[Any] = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] snake_case_ : Tuple = {} for i, token in enumerate(SCREAMING_SNAKE_CASE_ ): snake_case_ : Optional[int] = i snake_case_ : Optional[int] = WordpieceTokenizer(vocab=SCREAMING_SNAKE_CASE_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : str = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) snake_case_ : Optional[Any] = tokenizer.subword_tokenizer snake_case_ : List[str] = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) snake_case_ : str = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Dict = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) snake_case_ : Dict = tokenizer.encode('''ありがとう。''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) snake_case_ : List[str] = tokenizer.encode('''どういたしまして。''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) snake_case_ : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ ) snake_case_ : Any = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowerCAmelCase ( _UpperCAmelCase, unittest.TestCase ): lowerCamelCase_ : Optional[int] = BertJapaneseTokenizer lowerCamelCase_ : List[Any] = False def lowerCamelCase (self ) -> Dict: '''simple docstring''' super().setUp() snake_case_ : List[Any] = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] snake_case_ : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def lowerCamelCase (self , **__magic_name__ ) -> Optional[Any]: '''simple docstring''' return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **SCREAMING_SNAKE_CASE_ ) def lowerCamelCase (self , __magic_name__ ) -> Dict: '''simple docstring''' snake_case_ : Dict = 'こんにちは、世界。 \nこんばんは、世界。' snake_case_ : Tuple = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def lowerCamelCase (self ) -> Dict: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> Any: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> int: '''simple docstring''' pass # TODO add if relevant def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : Any = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) snake_case_ : Tuple = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] snake_case_ : Any = {} for i, token in enumerate(SCREAMING_SNAKE_CASE_ ): snake_case_ : Optional[int] = i snake_case_ : Any = CharacterTokenizer(vocab=SCREAMING_SNAKE_CASE_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : int = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) snake_case_ : int = tokenizer.encode('''ありがとう。''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) snake_case_ : str = tokenizer.encode('''どういたしまして。''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) snake_case_ : Optional[Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ ) snake_case_ : Tuple = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : List[str] = 'cl-tohoku/bert-base-japanese' snake_case_ : Tuple = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = 'cl-tohoku/bert-base-japanese' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) snake_case_ : Tuple = 'bert-base-cased' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
279
'''simple docstring''' import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class UpperCAmelCase_ (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = VQModel lowerCamelCase : Union[str, Any] = 'sample' @property def lowercase_ ( self , SCREAMING_SNAKE_CASE_=(32, 32) ) -> Any: __lowerCamelCase : Tuple = 4 __lowerCamelCase : Union[str, Any] = 3 __lowerCamelCase : List[Any] = floats_tensor((batch_size, num_channels) + sizes ).to(SCREAMING_SNAKE_CASE_ ) return {"sample": image} @property def lowercase_ ( self ) -> Optional[int]: return (3, 32, 32) @property def lowercase_ ( self ) -> List[Any]: return (3, 32, 32) def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Optional[Any] = { 'block_out_channels': [32, 64], 'in_channels': 3, 'out_channels': 3, 'down_block_types': ['DownEncoderBlock2D', 'DownEncoderBlock2D'], 'up_block_types': ['UpDecoderBlock2D', 'UpDecoderBlock2D'], 'latent_channels': 3, } __lowerCamelCase : Union[str, Any] = self.dummy_input return init_dict, inputs_dict def lowercase_ ( self ) -> str: pass def lowercase_ ( self ) -> Optional[int]: pass def lowercase_ ( self ) -> Any: __lowerCamelCase , __lowerCamelCase : Union[str, Any] = VQModel.from_pretrained('fusing/vqgan-dummy' , output_loading_info=SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) self.assertEqual(len(loading_info['missing_keys'] ) , 0 ) model.to(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def lowercase_ ( self ) -> int: __lowerCamelCase : List[Any] = VQModel.from_pretrained('fusing/vqgan-dummy' ) model.to(SCREAMING_SNAKE_CASE_ ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) __lowerCamelCase : List[str] = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) __lowerCamelCase : Optional[int] = image.to(SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): __lowerCamelCase : int = model(SCREAMING_SNAKE_CASE_ ).sample __lowerCamelCase : List[str] = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off __lowerCamelCase : Union[str, Any] = torch.tensor([-0.0_1_5_3, -0.4_0_4_4, -0.1_8_8_0, -0.5_1_6_1, -0.2_4_1_8, -0.4_0_7_2, -0.1_6_1_2, -0.0_6_3_3, -0.0_1_4_3] ) # fmt: on self.assertTrue(torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) )
185
0
'''simple docstring''' from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch SCREAMING_SNAKE_CASE_: Optional[int] =logging.get_logger(__name__) class __A ( UpperCamelCase__ ): a__ : Optional[Any] = ["""pixel_values"""] def __init__(self : int , __a : bool = True , __a : Optional[Dict[str, int]] = None , __a : PILImageResampling = PILImageResampling.BILINEAR , __a : bool = True , __a : Dict[str, int] = None , __a : bool = True , __a : Union[int, float] = 1 / 255 , __a : bool = True , __a : Optional[Union[float, List[float]]] = None , __a : Optional[Union[float, List[float]]] = None , **__a : Optional[Any] , ): super().__init__(**__a ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 256} UpperCAmelCase_ = get_size_dict(__a , default_to_square=__a ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(__a , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = resample UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def _lowercase (self : str , __a : np.ndarray , __a : Dict[str, int] , __a : PILImageResampling = PILImageResampling.BICUBIC , __a : Optional[Union[str, ChannelDimension]] = None , **__a : int , ): UpperCAmelCase_ = get_size_dict(__a , default_to_square=__a ) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase_ = get_resize_output_image_size(__a , size=size["shortest_edge"] , default_to_square=__a ) return resize(__a , size=__a , resample=__a , data_format=__a , **__a ) def _lowercase (self : str , __a : np.ndarray , __a : Dict[str, int] , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Dict , ): UpperCAmelCase_ = get_size_dict(__a ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(__a , size=(size["height"], size["width"]) , data_format=__a , **__a ) def _lowercase (self : Optional[Any] , __a : np.ndarray , __a : float , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Union[str, Any] ): return rescale(__a , scale=__a , data_format=__a , **__a ) def _lowercase (self : Tuple , __a : np.ndarray , __a : Union[float, List[float]] , __a : Union[float, List[float]] , __a : Optional[Union[str, ChannelDimension]] = None , **__a : List[Any] , ): return normalize(__a , mean=__a , std=__a , data_format=__a , **__a ) def _lowercase (self : Tuple , __a : ImageInput , __a : Optional[bool] = None , __a : Dict[str, int] = None , __a : PILImageResampling = None , __a : bool = None , __a : Dict[str, int] = None , __a : Optional[bool] = None , __a : Optional[float] = None , __a : Optional[bool] = None , __a : Optional[Union[float, List[float]]] = None , __a : Optional[Union[float, List[float]]] = None , __a : Optional[Union[str, TensorType]] = None , __a : Union[str, ChannelDimension] = ChannelDimension.FIRST , **__a : int , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(__a , default_to_square=__a ) UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(__a , param_name="crop_size" ) UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = make_list_of_images(__a ) if not valid_images(__a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = [to_numpy_array(__a ) for image in images] if do_resize: UpperCAmelCase_ = [self.resize(image=__a , size=__a , resample=__a ) for image in images] if do_center_crop: UpperCAmelCase_ = [self.center_crop(image=__a , size=__a ) for image in images] if do_rescale: UpperCAmelCase_ = [self.rescale(image=__a , scale=__a ) for image in images] if do_normalize: UpperCAmelCase_ = [self.normalize(image=__a , mean=__a , std=__a ) for image in images] UpperCAmelCase_ = [to_channel_dimension_format(__a , __a ) for image in images] UpperCAmelCase_ = {"pixel_values": images} return BatchFeature(data=__a , tensor_type=__a ) def _lowercase (self : Any , __a : Optional[int] , __a : List[Tuple] = None ): UpperCAmelCase_ = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(__a ) != len(__a ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(__a ): UpperCAmelCase_ = target_sizes.numpy() UpperCAmelCase_ = [] for idx in range(len(__a ) ): UpperCAmelCase_ = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=__a ) UpperCAmelCase_ = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(__a ) else: UpperCAmelCase_ = logits.argmax(dim=1 ) UpperCAmelCase_ = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
352
'''simple docstring''' import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch SCREAMING_SNAKE_CASE_: Dict =True except ImportError: SCREAMING_SNAKE_CASE_: str =False try: from torch.hub import _get_torch_home SCREAMING_SNAKE_CASE_: Optional[Any] =_get_torch_home() except ImportError: SCREAMING_SNAKE_CASE_: Union[str, Any] =os.path.expanduser( os.getenv('TORCH_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'torch')) ) SCREAMING_SNAKE_CASE_: int =os.path.join(torch_cache_home, 'transformers') SCREAMING_SNAKE_CASE_: Tuple ='https://cdn.huggingface.co' SCREAMING_SNAKE_CASE_: str ='https://s3.amazonaws.com/models.huggingface.co/bert' SCREAMING_SNAKE_CASE_: str ='/'.join(str(Path(__file__).resolve()).split('/')[:-1]) SCREAMING_SNAKE_CASE_: Optional[Any] =os.path.join(PATH, 'config.yaml') SCREAMING_SNAKE_CASE_: Optional[Any] =os.path.join(PATH, 'attributes.txt') SCREAMING_SNAKE_CASE_: Any =os.path.join(PATH, 'objects.txt') SCREAMING_SNAKE_CASE_: Optional[int] =os.getenv('PYTORCH_PRETRAINED_BERT_CACHE', default_cache_path) SCREAMING_SNAKE_CASE_: int =os.getenv('PYTORCH_TRANSFORMERS_CACHE', PYTORCH_PRETRAINED_BERT_CACHE) SCREAMING_SNAKE_CASE_: List[str] =os.getenv('TRANSFORMERS_CACHE', PYTORCH_TRANSFORMERS_CACHE) SCREAMING_SNAKE_CASE_: str ='pytorch_model.bin' SCREAMING_SNAKE_CASE_: Dict ='config.yaml' def lowerCAmelCase_ ( snake_case_ : Optional[int]=OBJECTS , snake_case_ : Optional[Any]=ATTRIBUTES ) -> Any: '''simple docstring''' UpperCAmelCase_ = [] with open(snake_case_ ) as f: for object in f.readlines(): vg_classes.append(object.split("," )[0].lower().strip() ) UpperCAmelCase_ = [] with open(snake_case_ ) as f: for object in f.readlines(): vg_attrs.append(object.split("," )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCAmelCase_ ( snake_case_ : Optional[int] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = OrderedDict() with open(snake_case_ , "rb" ) as f: UpperCAmelCase_ = pkl.load(snake_case_ )["model"] for k in copy.deepcopy(list(ckp.keys() ) ): UpperCAmelCase_ = ckp.pop(snake_case_ ) if isinstance(snake_case_ , np.ndarray ): UpperCAmelCase_ = torch.tensor(snake_case_ ) else: assert isinstance(snake_case_ , torch.tensor ), type(snake_case_ ) UpperCAmelCase_ = v return r class __A : a__ : Optional[Any] = {} def __init__(self : Union[str, Any] , __a : dict , __a : str = "root" , __a : str=0 ): UpperCAmelCase_ = name UpperCAmelCase_ = level UpperCAmelCase_ = {} for k, v in dictionary.items(): if v is None: raise ValueError() UpperCAmelCase_ = copy.deepcopy(__a ) UpperCAmelCase_ = copy.deepcopy(__a ) if isinstance(__a , __a ): UpperCAmelCase_ = Config(__a , name=__a , level=level + 1 ) UpperCAmelCase_ = v setattr(self , __a , __a ) UpperCAmelCase_ = d def __repr__(self : List[Any] ): return str(list((self._pointer.keys()) ) ) def __setattr__(self : int , __a : str , __a : Dict ): UpperCAmelCase_ = val UpperCAmelCase_ = val UpperCAmelCase_ = key.split("." ) UpperCAmelCase_ = len(__a ) - 1 UpperCAmelCase_ = self._pointer if len(__a ) > 1: for i, l in enumerate(__a ): if hasattr(self , __a ) and isinstance(getattr(self , __a ) , __a ): setattr(getattr(self , __a ) , ".".join(levels[i:] ) , __a ) if l == last_level: UpperCAmelCase_ = val else: UpperCAmelCase_ = pointer[l] def _lowercase (self : Optional[Any] ): return self._pointer def _lowercase (self : int , __a : Union[str, Any] , __a : str ): with open(f"""{file_name}""" , "w" ) as stream: dump(__a , __a ) def _lowercase (self : Any , __a : Optional[Any] , __a : List[str] ): with open(f"""{file_name}""" , "w" ) as stream: json.dump(__a , __a ) @staticmethod def _lowercase (__a : str ): with open(__a ) as stream: UpperCAmelCase_ = load(__a , Loader=__a ) return data def __str__(self : Dict ): UpperCAmelCase_ = " " if self._name != "root": UpperCAmelCase_ = f"""{t * (self._level-1)}{self._name}:\n""" else: UpperCAmelCase_ = "" UpperCAmelCase_ = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__a , __a ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__a ).__name__})\n""" UpperCAmelCase_ = level return r[:-1] @classmethod def _lowercase (cls : Tuple , __a : str , **__a : Dict ): UpperCAmelCase_ , UpperCAmelCase_ = cls.get_config_dict(__a , **__a ) return cls(__a ) @classmethod def _lowercase (cls : Any , __a : str , **__a : Dict ): UpperCAmelCase_ = kwargs.pop("cache_dir" , __a ) UpperCAmelCase_ = kwargs.pop("force_download" , __a ) UpperCAmelCase_ = kwargs.pop("resume_download" , __a ) UpperCAmelCase_ = kwargs.pop("proxies" , __a ) UpperCAmelCase_ = kwargs.pop("local_files_only" , __a ) if os.path.isdir(__a ): UpperCAmelCase_ = os.path.join(__a , __a ) elif os.path.isfile(__a ) or is_remote_url(__a ): UpperCAmelCase_ = pretrained_model_name_or_path else: UpperCAmelCase_ = hf_bucket_url(__a , filename=__a , use_cdn=__a ) try: # Load from URL or cache if already cached UpperCAmelCase_ = cached_path( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , ) # Load config dict if resolved_config_file is None: raise EnvironmentError UpperCAmelCase_ = Config.load_yaml(__a ) except EnvironmentError: UpperCAmelCase_ = "Can't load config for" raise EnvironmentError(__a ) if resolved_config_file == config_file: print("loading configuration file from path" ) else: print("loading configuration file cache" ) return Config.load_yaml(__a ), kwargs def lowerCAmelCase_ ( snake_case_ : str ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = torch.load("dump.pt" , map_location=in_tensor.device ) UpperCAmelCase_ = in_tensor.numpy() UpperCAmelCase_ = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(snake_case_ , snake_case_ , rtol=0.01 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(snake_case_ , snake_case_ , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*1_00:.4f} %""" " element-wise mismatch" ) raise Exception("tensors are all good" ) # Hugging face functions below def lowerCAmelCase_ ( snake_case_ : Optional[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = urlparse(snake_case_ ) return parsed.scheme in ("http", "https") def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : str , snake_case_ : Optional[int]=True ) -> str: '''simple docstring''' UpperCAmelCase_ = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX UpperCAmelCase_ = "/" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Union[str, Any] , snake_case_ : Optional[int]=None , snake_case_ : List[Any]=0 , snake_case_ : int=None , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = "python/{}".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(snake_case_ , snake_case_ ): ua += "; " + "; ".join("{}/{}".format(snake_case_ , snake_case_ ) for k, v in user_agent.items() ) elif isinstance(snake_case_ , snake_case_ ): ua += "; " + user_agent UpperCAmelCase_ = {"user-agent": ua} if resume_size > 0: UpperCAmelCase_ = "bytes=%d-" % (resume_size,) UpperCAmelCase_ = requests.get(snake_case_ , stream=snake_case_ , proxies=snake_case_ , headers=snake_case_ ) if response.status_code == 4_16: # Range not satisfiable return UpperCAmelCase_ = response.headers.get("Content-Length" ) UpperCAmelCase_ = resume_size + int(snake_case_ ) if content_length is not None else None UpperCAmelCase_ = tqdm( unit="B" , unit_scale=snake_case_ , total=snake_case_ , initial=snake_case_ , desc="Downloading" , ) for chunk in response.iter_content(chunk_size=10_24 ): if chunk: # filter out keep-alive new chunks progress.update(len(snake_case_ ) ) temp_file.write(snake_case_ ) progress.close() def lowerCAmelCase_ ( snake_case_ : Optional[int] , snake_case_ : str=None , snake_case_ : List[str]=False , snake_case_ : List[str]=None , snake_case_ : int=10 , snake_case_ : Any=False , snake_case_ : int=None , snake_case_ : str=False , ) -> str: '''simple docstring''' if cache_dir is None: UpperCAmelCase_ = TRANSFORMERS_CACHE if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) UpperCAmelCase_ = None if not local_files_only: try: UpperCAmelCase_ = requests.head(snake_case_ , allow_redirects=snake_case_ , proxies=snake_case_ , timeout=snake_case_ ) if response.status_code == 2_00: UpperCAmelCase_ = response.headers.get("ETag" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass UpperCAmelCase_ = url_to_filename(snake_case_ , snake_case_ ) # get cache path to put the file UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(snake_case_ ): return cache_path else: UpperCAmelCase_ = [ file for file in fnmatch.filter(os.listdir(snake_case_ ) , filename + ".*" ) if not file.endswith(".json" ) and not file.endswith(".lock" ) ] if len(snake_case_ ) > 0: return os.path.join(snake_case_ , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( "Cannot find the requested files in the cached path and outgoing traffic has been" " disabled. To enable model look-ups and downloads online, set 'local_files_only'" " to False." ) return None # From now on, etag is not None. if os.path.exists(snake_case_ ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. UpperCAmelCase_ = cache_path + ".lock" with FileLock(snake_case_ ): # If the download just completed while the lock was activated. if os.path.exists(snake_case_ ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: UpperCAmelCase_ = cache_path + ".incomplete" @contextmanager def _resumable_file_manager(): with open(snake_case_ , "a+b" ) as f: yield f UpperCAmelCase_ = _resumable_file_manager if os.path.exists(snake_case_ ): UpperCAmelCase_ = os.stat(snake_case_ ).st_size else: UpperCAmelCase_ = 0 else: UpperCAmelCase_ = partial(tempfile.NamedTemporaryFile , dir=snake_case_ , delete=snake_case_ ) UpperCAmelCase_ = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( "%s not found in cache or force_download set to True, downloading to %s" , snake_case_ , temp_file.name , ) http_get( snake_case_ , snake_case_ , proxies=snake_case_ , resume_size=snake_case_ , user_agent=snake_case_ , ) os.replace(temp_file.name , snake_case_ ) UpperCAmelCase_ = {"url": url, "etag": etag} UpperCAmelCase_ = cache_path + ".json" with open(snake_case_ , "w" ) as meta_file: json.dump(snake_case_ , snake_case_ ) return cache_path def lowerCAmelCase_ ( snake_case_ : Optional[Any] , snake_case_ : Any=None ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = url.encode("utf-8" ) UpperCAmelCase_ = shaaaa(snake_case_ ) UpperCAmelCase_ = url_hash.hexdigest() if etag: UpperCAmelCase_ = etag.encode("utf-8" ) UpperCAmelCase_ = shaaaa(snake_case_ ) filename += "." + etag_hash.hexdigest() if url.endswith(".h5" ): filename += ".h5" return filename def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : Tuple=None , snake_case_ : int=False , snake_case_ : Any=None , snake_case_ : List[Any]=False , snake_case_ : Any=None , snake_case_ : Any=False , snake_case_ : List[str]=False , snake_case_ : str=False , ) -> Union[str, Any]: '''simple docstring''' if cache_dir is None: UpperCAmelCase_ = TRANSFORMERS_CACHE if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) if isinstance(snake_case_ , snake_case_ ): UpperCAmelCase_ = str(snake_case_ ) if is_remote_url(snake_case_ ): # URL, so get it from the cache (downloading if necessary) UpperCAmelCase_ = get_from_cache( snake_case_ , cache_dir=snake_case_ , force_download=snake_case_ , proxies=snake_case_ , resume_download=snake_case_ , user_agent=snake_case_ , local_files_only=snake_case_ , ) elif os.path.exists(snake_case_ ): # File, and it exists. UpperCAmelCase_ = url_or_filename elif urlparse(snake_case_ ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("file {} not found".format(snake_case_ ) ) else: # Something unknown raise ValueError("unable to parse {} as a URL or as a local path".format(snake_case_ ) ) if extract_compressed_file: if not is_zipfile(snake_case_ ) and not tarfile.is_tarfile(snake_case_ ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" UpperCAmelCase_ , UpperCAmelCase_ = os.path.split(snake_case_ ) UpperCAmelCase_ = output_file.replace("." , "-" ) + "-extracted" UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) if os.path.isdir(snake_case_ ) and os.listdir(snake_case_ ) and not force_extract: return output_path_extracted # Prevent parallel extractions UpperCAmelCase_ = output_path + ".lock" with FileLock(snake_case_ ): shutil.rmtree(snake_case_ , ignore_errors=snake_case_ ) os.makedirs(snake_case_ ) if is_zipfile(snake_case_ ): with ZipFile(snake_case_ , "r" ) as zip_file: zip_file.extractall(snake_case_ ) zip_file.close() elif tarfile.is_tarfile(snake_case_ ): UpperCAmelCase_ = tarfile.open(snake_case_ ) tar_file.extractall(snake_case_ ) tar_file.close() else: raise EnvironmentError("Archive format of {} could not be identified".format(snake_case_ ) ) return output_path_extracted return output_path def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : Optional[int]="," ) -> int: '''simple docstring''' assert isinstance(snake_case_ , snake_case_ ) if os.path.isfile(snake_case_ ): with open(snake_case_ ) as f: UpperCAmelCase_ = eval(f.read() ) else: UpperCAmelCase_ = requests.get(snake_case_ ) try: UpperCAmelCase_ = requests.json() except Exception: UpperCAmelCase_ = req.content.decode() assert data is not None, "could not connect" try: UpperCAmelCase_ = eval(snake_case_ ) except Exception: UpperCAmelCase_ = data.split("\n" ) req.close() return data def lowerCAmelCase_ ( snake_case_ : List[str] ) -> Any: '''simple docstring''' UpperCAmelCase_ = requests.get(snake_case_ ) UpperCAmelCase_ = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCAmelCase_ ( snake_case_ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = url.split("/" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(snake_case_ ) with open(snake_case_ , "rb" ) as stream: UpperCAmelCase_ = pkl.load(snake_case_ ) UpperCAmelCase_ = weights.pop("model" ) UpperCAmelCase_ = {} for k, v in model.items(): UpperCAmelCase_ = torch.from_numpy(snake_case_ ) if "running_var" in k: UpperCAmelCase_ = torch.tensor([0] ) UpperCAmelCase_ = k.replace("running_var" , "num_batches_tracked" ) UpperCAmelCase_ = zero return new def lowerCAmelCase_ ( ) -> int: '''simple docstring''' print(f"""{os.path.abspath(os.path.join(snake_case_ , os.pardir ) )}/demo.ipynb""" ) def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : Any="RGB" ) -> Dict: '''simple docstring''' assert isinstance(snake_case_ , snake_case_ ) if os.path.isfile(snake_case_ ): UpperCAmelCase_ = cva.imread(snake_case_ ) else: UpperCAmelCase_ = get_image_from_url(snake_case_ ) assert img is not None, f"""could not connect to: {im}""" UpperCAmelCase_ = cva.cvtColor(snake_case_ , cva.COLOR_BGR2RGB ) if input_format == "RGB": UpperCAmelCase_ = img[:, :, ::-1] return img def lowerCAmelCase_ ( snake_case_ : Tuple , snake_case_ : Union[str, Any]=1 ) -> str: '''simple docstring''' return (images[i : i + batch] for i in range(0 , len(snake_case_ ) , snake_case_ ))
106
0
"""simple docstring""" import unittest import torch from torch import nn from diffusers.models.activations import get_activation class lowerCAmelCase_ ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" snake_case = get_activation('swish' ) self.assertIsInstance(lowerCAmelCase , nn.SiLU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def snake_case ( self ): """simple docstring""" snake_case = get_activation('silu' ) self.assertIsInstance(lowerCAmelCase , nn.SiLU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def snake_case ( self ): """simple docstring""" snake_case = get_activation('mish' ) self.assertIsInstance(lowerCAmelCase , nn.Mish ) self.assertEqual(act(torch.tensor(-2_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def snake_case ( self ): """simple docstring""" snake_case = get_activation('gelu' ) self.assertIsInstance(lowerCAmelCase , nn.GELU ) self.assertEqual(act(torch.tensor(-1_00 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 )
150
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class lowerCAmelCase_ : """simple docstring""" # setable values _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : Optional[jnp.ndarray] = None _lowerCAmelCase : Optional[jnp.ndarray] = None # sigma(t_i) @classmethod def snake_case ( cls ): """simple docstring""" return cls() @dataclass class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase : jnp.ndarray _lowerCAmelCase : jnp.ndarray _lowerCAmelCase : KarrasVeSchedulerState class lowerCAmelCase_ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" @property def snake_case ( self ): """simple docstring""" return True @register_to_config def __init__( self , lowerCAmelCase = 0.02 , lowerCAmelCase = 1_00 , lowerCAmelCase = 1.0_07 , lowerCAmelCase = 80 , lowerCAmelCase = 0.05 , lowerCAmelCase = 50 , ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" return KarrasVeSchedulerState.create() def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = () ): """simple docstring""" snake_case = jnp.arange(0 , lowerCAmelCase )[::-1].copy() snake_case = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=lowerCAmelCase , schedule=jnp.array(lowerCAmelCase , dtype=jnp.floataa ) , timesteps=lowerCAmelCase , ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): """simple docstring""" if self.config.s_min <= sigma <= self.config.s_max: snake_case = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: snake_case = 0 # sample eps ~ N(0, S_noise^2 * I) snake_case = random.split(lowerCAmelCase , num=1 ) snake_case = self.config.s_noise * random.normal(key=lowerCAmelCase , shape=sample.shape ) snake_case = sigma + gamma * sigma snake_case = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = True , ): """simple docstring""" snake_case = sample_hat + sigma_hat * model_output snake_case = (sample_hat - pred_original_sample) / sigma_hat snake_case = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowerCAmelCase , derivative=lowerCAmelCase , state=lowerCAmelCase ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = True , ): """simple docstring""" snake_case = sample_prev + sigma_prev * model_output snake_case = (sample_prev - pred_original_sample) / sigma_prev snake_case = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowerCAmelCase , derivative=lowerCAmelCase , state=lowerCAmelCase ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" raise NotImplementedError()
150
1
"""simple docstring""" def snake_case (A_ :Union[str, Any] ) -> int: '''simple docstring''' if not head: return True # split the list to two parts a : Dict = head.next, head while fast and fast.next: a : str = fast.next.next a : str = slow.next a : Optional[Any] = slow.next a : List[Any] = None # Don't forget here! But forget still works! # reverse the second part a : Dict = None while second: a : int = second.next a : int = node a : List[Any] = second a : Any = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False a : str = node.next a : str = head.next return True def snake_case (A_ :List[str] ) -> Tuple: '''simple docstring''' if not head or not head.next: return True # 1. Get the midpoint (slow) a : Any = head while fast and fast.next: a : Dict = fast.next.next, slow.next # 2. Push the second half into the stack a : Union[str, Any] = [slow.val] while slow.next: a : str = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False a : List[str] = cur.next return True def snake_case (A_ :Any ) -> Tuple: '''simple docstring''' if not head or not head.next: return True a : List[str] = {} a : List[Any] = 0 while head: if head.val in d: d[head.val].append(A_ ) else: a : Dict = [pos] a : Optional[Any] = head.next pos += 1 a : Dict = pos - 1 a : Union[str, Any] = 0 for v in d.values(): if len(A_ ) % 2 != 0: middle += 1 else: a : Any = 0 for i in range(0 , len(A_ ) ): if v[i] + v[len(A_ ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
365
"""simple docstring""" def snake_case (A_ :int ): '''simple docstring''' if isinstance(A_ , A_ ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(A_ , A_ ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" a : List[Any] = False if num < 0: a : Optional[int] = True a : Dict = -num a : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(A_ ) for e in binary ) return "0b" + "".join(str(A_ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
186
0
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __A = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" __A = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" __A = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/google-research/google-research/tree/master/rouge'] , reference_urls=[ 'https://en.wikipedia.org/wiki/ROUGE_(metric)', 'https://github.com/google-research/google-research/tree/master/rouge', ] , ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=False ) -> List[str]: '''simple docstring''' if rouge_types is None: __lowerCamelCase = ["rouge1", "rouge2", "rougeL", "rougeLsum"] __lowerCamelCase = rouge_scorer.RougeScorer(rouge_types=__a , use_stemmer=__a ) if use_aggregator: __lowerCamelCase = scoring.BootstrapAggregator() else: __lowerCamelCase = [] for ref, pred in zip(__a , __a ): __lowerCamelCase = scorer.score(__a , __a ) if use_aggregator: aggregator.add_scores(__a ) else: scores.append(__a ) if use_aggregator: __lowerCamelCase = aggregator.aggregate() else: __lowerCamelCase = {} for key in scores[0]: __lowerCamelCase = [score[key] for score in scores] return result
90
from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( 'The RoBERTa Model transformer with early exiting (DeeRoBERTa). ' , a , ) class UpperCAmelCase_ ( a): lowerCamelCase__ = RobertaConfig lowerCamelCase__ = 'roberta' def __init__( self, __a): '''simple docstring''' super().__init__(__a) _lowerCAmelCase : Optional[Any] = RobertaEmbeddings(__a) self.init_weights() @add_start_docstrings( 'RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top,\n also takes care of multi-layer training. ' , a , ) class UpperCAmelCase_ ( a): lowerCamelCase__ = RobertaConfig lowerCamelCase__ = 'roberta' def __init__( self, __a): '''simple docstring''' super().__init__(__a) _lowerCAmelCase : Optional[int] = config.num_labels _lowerCAmelCase : Optional[int] = config.num_hidden_layers _lowerCAmelCase : Optional[int] = DeeRobertaModel(__a) _lowerCAmelCase : Union[str, Any] = nn.Dropout(config.hidden_dropout_prob) _lowerCAmelCase : List[str] = nn.Linear(config.hidden_size, self.config.num_labels) @add_start_docstrings_to_model_forward(__a) def snake_case__ ( self, __a=None, __a=None, __a=None, __a=None, __a=None, __a=None, __a=None, __a=-1, __a=False, ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.num_layers try: _lowerCAmelCase : List[Any] = self.roberta( __a, attention_mask=__a, token_type_ids=__a, position_ids=__a, head_mask=__a, inputs_embeds=__a, ) _lowerCAmelCase : List[Any] = outputs[1] _lowerCAmelCase : Dict = self.dropout(__a) _lowerCAmelCase : Dict = self.classifier(__a) _lowerCAmelCase : Optional[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: _lowerCAmelCase : Tuple = e.message _lowerCAmelCase : Union[str, Any] = e.exit_layer _lowerCAmelCase : List[Any] = outputs[0] if not self.training: _lowerCAmelCase : int = entropy(__a) _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : str = [] if labels is not None: if self.num_labels == 1: # We are doing regression _lowerCAmelCase : Optional[Any] = MSELoss() _lowerCAmelCase : int = loss_fct(logits.view(-1), labels.view(-1)) else: _lowerCAmelCase : Optional[Any] = CrossEntropyLoss() _lowerCAmelCase : Optional[Any] = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) # work with highway exits _lowerCAmelCase : Optional[int] = [] for highway_exit in outputs[-1]: _lowerCAmelCase : Any = highway_exit[0] if not self.training: highway_logits_all.append(__a) highway_entropy.append(highway_exit[2]) if self.num_labels == 1: # We are doing regression _lowerCAmelCase : List[str] = MSELoss() _lowerCAmelCase : List[Any] = loss_fct(highway_logits.view(-1), labels.view(-1)) else: _lowerCAmelCase : Dict = CrossEntropyLoss() _lowerCAmelCase : Optional[Any] = loss_fct(highway_logits.view(-1, self.num_labels), labels.view(-1)) highway_losses.append(__a) if train_highway: _lowerCAmelCase : int = (sum(highway_losses[:-1]),) + outputs # exclude the final highway, of course else: _lowerCAmelCase : Any = (loss,) + outputs if not self.training: _lowerCAmelCase : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: _lowerCAmelCase : Optional[Any] = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
36
0
"""simple docstring""" import sys __UpperCAmelCase = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def _snake_case ( lowercase__ : str ) -> int: '''simple docstring''' lowerCAmelCase_ :Union[str, Any] = 1 for digit in s: product *= int(lowercase__ ) return product def _snake_case ( lowercase__ : str = N ) -> int: '''simple docstring''' lowerCAmelCase_ :Dict = -sys.maxsize - 1 lowerCAmelCase_ :Optional[int] = n[:1_3] lowerCAmelCase_ :Any = 1_3 while cur_index < len(lowercase__ ) - 1_3: if int(n[cur_index] ) >= int(substr[0] ): lowerCAmelCase_ :Dict = substr[1:] + n[cur_index] cur_index += 1 else: lowerCAmelCase_ :List[Any] = max(lowercase__ , str_eval(lowercase__ ) ) lowerCAmelCase_ :Optional[Any] = n[cur_index : cur_index + 1_3] cur_index += 1_3 return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
1
"""simple docstring""" import os from math import logaa def _snake_case ( lowercase__ : str = "base_exp.txt" ) -> int: '''simple docstring''' lowerCAmelCase_ :float = 0 lowerCAmelCase_ :Union[str, Any] = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(lowercase__ ) , lowercase__ ) ) ): lowerCAmelCase_ , lowerCAmelCase_ :Union[str, Any] = list(map(lowercase__ , line.split(""",""" ) ) ) if x * logaa(lowercase__ ) > largest: lowerCAmelCase_ :Any = x * logaa(lowercase__ ) lowerCAmelCase_ :List[Any] = i + 1 return result if __name__ == "__main__": print(solution())
1
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class _SCREAMING_SNAKE_CASE : def __init__( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : List[str]=True , __lowerCamelCase : Dict=False , __lowerCamelCase : Any=10 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : Optional[Any]=32 * 8 , __lowerCamelCase : Tuple=32 * 8 , __lowerCamelCase : Union[str, Any]=4 , __lowerCamelCase : Tuple=64 , ): UpperCamelCase :int = parent UpperCamelCase :Optional[Any] = batch_size UpperCamelCase :Tuple = is_training UpperCamelCase :Tuple = use_auxiliary_loss UpperCamelCase :Dict = num_queries UpperCamelCase :Any = num_channels UpperCamelCase :List[Any] = min_size UpperCamelCase :Any = max_size UpperCamelCase :List[str] = num_labels UpperCamelCase :Dict = hidden_dim UpperCamelCase :Optional[Any] = hidden_dim def _A ( self : Any ): UpperCamelCase :Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( __lowerCamelCase ) UpperCamelCase :Optional[int] = torch.ones([self.batch_size, self.min_size, self.max_size] , device=__lowerCamelCase ) UpperCamelCase :Optional[Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=__lowerCamelCase ) > 0.5 ).float() UpperCamelCase :Any = (torch.rand((self.batch_size, self.num_labels) , device=__lowerCamelCase ) > 0.5).long() UpperCamelCase :List[Any] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _A ( self : Optional[int] ): UpperCamelCase :int = MaskaFormerConfig( hidden_size=self.hidden_dim , ) UpperCamelCase :Tuple = self.num_queries UpperCamelCase :Optional[Any] = self.num_labels UpperCamelCase :Any = [1, 1, 1, 1] UpperCamelCase :List[str] = self.num_channels UpperCamelCase :int = 64 UpperCamelCase :Union[str, Any] = 128 UpperCamelCase :Optional[Any] = self.hidden_dim UpperCamelCase :List[Any] = self.hidden_dim UpperCamelCase :List[str] = self.hidden_dim return config def _A ( self : List[Any] ): UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase :List[str] = self.prepare_config_and_inputs() UpperCamelCase :List[Any] = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def _A ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ): UpperCamelCase :Optional[Any] = output.encoder_hidden_states UpperCamelCase :Optional[Any] = output.pixel_decoder_hidden_states UpperCamelCase :Dict = output.transformer_decoder_hidden_states self.parent.assertTrue(len(__lowerCamelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__lowerCamelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__lowerCamelCase ) , config.decoder_layers ) def _A ( self : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Optional[Any]=False ): with torch.no_grad(): UpperCamelCase :Optional[Any] = MaskaFormerModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :List[str] = model(pixel_values=__lowerCamelCase , pixel_mask=__lowerCamelCase ) UpperCamelCase :Tuple = model(__lowerCamelCase , output_hidden_states=__lowerCamelCase ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(__lowerCamelCase , __lowerCamelCase ) def _A ( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : int ): UpperCamelCase :Optional[int] = MaskaFormerForUniversalSegmentation(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() def comm_check_on_output(__lowerCamelCase : Optional[Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCamelCase :Optional[int] = model(pixel_values=__lowerCamelCase , pixel_mask=__lowerCamelCase ) UpperCamelCase :Any = model(__lowerCamelCase ) comm_check_on_output(__lowerCamelCase ) UpperCamelCase :Tuple = model( pixel_values=__lowerCamelCase , pixel_mask=__lowerCamelCase , mask_labels=__lowerCamelCase , class_labels=__lowerCamelCase ) comm_check_on_output(__lowerCamelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case__ : Dict = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () snake_case__ : List[Any] = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} snake_case__ : List[str] = False snake_case__ : Any = False snake_case__ : Optional[int] = False snake_case__ : Tuple = False def _A ( self : List[str] ): UpperCamelCase :str = MaskaFormerModelTester(self ) UpperCamelCase :Union[str, Any] = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def _A ( self : Optional[Any] ): self.config_tester.run_common_tests() def _A ( self : str ): UpperCamelCase , UpperCamelCase :List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(__lowerCamelCase , **__lowerCamelCase , output_hidden_states=__lowerCamelCase ) def _A ( self : Union[str, Any] ): UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*__lowerCamelCase ) @unittest.skip(reason="""Mask2Former does not use inputs_embeds""" ) def _A ( self : List[Any] ): pass @unittest.skip(reason="""Mask2Former does not have a get_input_embeddings method""" ) def _A ( self : str ): pass @unittest.skip(reason="""Mask2Former is not a generative model""" ) def _A ( self : str ): pass @unittest.skip(reason="""Mask2Former does not use token embeddings""" ) def _A ( self : Optional[Any] ): pass @require_torch_multi_gpu @unittest.skip( reason="""Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def _A ( self : Optional[Any] ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _A ( self : Dict ): pass def _A ( self : str ): UpperCamelCase , UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :List[Any] = model_class(__lowerCamelCase ) UpperCamelCase :str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase :Dict = [*signature.parameters.keys()] UpperCamelCase :Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) @slow def _A ( self : int ): for model_name in ["facebook/mask2former-swin-small-coco-instance"]: UpperCamelCase :int = MaskaFormerModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _A ( self : List[str] ): UpperCamelCase :Dict = (self.model_tester.min_size,) * 2 UpperCamelCase :Tuple = { """pixel_values""": torch.randn((2, 3, *size) , device=__lowerCamelCase ), """mask_labels""": torch.randn((2, 10, *size) , device=__lowerCamelCase ), """class_labels""": torch.zeros(2 , 10 , device=__lowerCamelCase ).long(), } UpperCamelCase :str = self.model_tester.get_config() UpperCamelCase :str = MaskaFormerForUniversalSegmentation(__lowerCamelCase ).to(__lowerCamelCase ) UpperCamelCase :Optional[Any] = model(**__lowerCamelCase ) self.assertTrue(outputs.loss is not None ) def _A ( self : Dict ): UpperCamelCase , UpperCamelCase :str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(__lowerCamelCase , **__lowerCamelCase , output_hidden_states=__lowerCamelCase ) def _A ( self : Dict ): UpperCamelCase , UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :Optional[int] = model_class(__lowerCamelCase ).to(__lowerCamelCase ) UpperCamelCase :Union[str, Any] = model(**__lowerCamelCase , output_attentions=__lowerCamelCase ) self.assertTrue(outputs.attentions is not None ) def _A ( self : List[str] ): if not self.model_tester.is_training: return UpperCamelCase :Dict = self.all_model_classes[1] UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs() UpperCamelCase :Optional[Any] = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() UpperCamelCase :int = model(__lowerCamelCase , mask_labels=__lowerCamelCase , class_labels=__lowerCamelCase ).loss loss.backward() def _A ( self : List[str] ): UpperCamelCase :Optional[int] = self.all_model_classes[1] UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs() UpperCamelCase :List[Any] = True UpperCamelCase :str = True UpperCamelCase :int = model_class(__lowerCamelCase ).to(__lowerCamelCase ) model.train() UpperCamelCase :Union[str, Any] = model(__lowerCamelCase , mask_labels=__lowerCamelCase , class_labels=__lowerCamelCase ) UpperCamelCase :Optional[Any] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCamelCase :Any = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() UpperCamelCase :Optional[int] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCamelCase :Any = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=__lowerCamelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) UpperCAmelCase_ : Optional[Any] = 1E-4 def SCREAMING_SNAKE_CASE_ ( ) -> Any: """simple docstring""" UpperCamelCase :Union[str, Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _A ( self : Tuple ): return "facebook/mask2former-swin-small-coco-instance" @cached_property def _A ( self : Dict ): return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def _A ( self : Tuple ): UpperCamelCase :List[Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(__lowerCamelCase ) UpperCamelCase :Optional[int] = self.default_image_processor UpperCamelCase :Optional[Any] = prepare_img() UpperCamelCase :Optional[Any] = image_processor(__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) UpperCamelCase :str = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(__lowerCamelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCamelCase :Optional[int] = model(**__lowerCamelCase ) UpperCamelCase :List[str] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(__lowerCamelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) UpperCamelCase :Tuple = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(__lowerCamelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) UpperCamelCase :Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(__lowerCamelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) def _A ( self : Any ): UpperCamelCase :Dict = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(__lowerCamelCase ).eval() UpperCamelCase :str = self.default_image_processor UpperCamelCase :Dict = prepare_img() UpperCamelCase :str = image_processor(__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) UpperCamelCase :str = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(__lowerCamelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCamelCase :List[Any] = model(**__lowerCamelCase ) # masks_queries_logits UpperCamelCase :int = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) UpperCamelCase :Any = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] UpperCamelCase :Optional[int] = torch.tensor(__lowerCamelCase ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) # class_queries_logits UpperCamelCase :List[Any] = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) UpperCamelCase :Union[str, Any] = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) def _A ( self : Tuple ): UpperCamelCase :List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(__lowerCamelCase ).eval() UpperCamelCase :int = self.default_image_processor UpperCamelCase :List[str] = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) UpperCamelCase :Optional[Any] = inputs["""pixel_values"""].to(__lowerCamelCase ) UpperCamelCase :Optional[Any] = [el.to(__lowerCamelCase ) for el in inputs["""mask_labels"""]] UpperCamelCase :Any = [el.to(__lowerCamelCase ) for el in inputs["""class_labels"""]] with torch.no_grad(): UpperCamelCase :List[str] = model(**__lowerCamelCase ) self.assertTrue(outputs.loss is not None )
38
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Tuple = logging.get_logger(__name__) lowerCAmelCase__ : Union[str, Any] = {'''vocab_file''': '''sentencepiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ : Dict = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, '''tokenizer_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ : List[Any] = { '''google/rembert''': 2_56, } lowerCAmelCase__ : List[str] = '''▁''' class __snake_case ( _lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RemBertTokenizer def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase="[CLS]" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<unk>" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<pad>" , __UpperCamelCase="[CLS]" , __UpperCamelCase="[MASK]" , **__UpperCamelCase , ) -> Tuple: '''simple docstring''' snake_case__ : Optional[Any] = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token super().__init__( __UpperCamelCase , tokenizer_file=__UpperCamelCase , do_lower_case=__UpperCamelCase , remove_space=__UpperCamelCase , keep_accents=__UpperCamelCase , bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , pad_token=__UpperCamelCase , cls_token=__UpperCamelCase , mask_token=__UpperCamelCase , **__UpperCamelCase , ) snake_case__ : int = do_lower_case snake_case__ : Any = remove_space snake_case__ : List[Any] = keep_accents snake_case__ : Dict = vocab_file snake_case__ : int = False if not self.vocab_file else True def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : Dict = [self.sep_token_id] snake_case__ : List[str] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__UpperCamelCase )) + [1] + ([0] * len(__UpperCamelCase )) + [1] return [1] + ([0] * len(__UpperCamelCase )) + [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : List[Any] = [self.sep_token_id] snake_case__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(__UpperCamelCase ): logger.error('Vocabulary path ({}) should be a directory'.format(__UpperCamelCase ) ) return snake_case__ : List[str] = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ): copyfile(self.vocab_file , __UpperCamelCase ) return (out_vocab_file,)
143
0
"""simple docstring""" from __future__ import annotations import time __SCREAMING_SNAKE_CASE : Any = list[tuple[int, int]] __SCREAMING_SNAKE_CASE : Union[str, Any] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __SCREAMING_SNAKE_CASE : Any = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pos_x _lowerCamelCase = pos_y _lowerCamelCase = (pos_y, pos_x) _lowerCamelCase = goal_x _lowerCamelCase = goal_y _lowerCamelCase = parent class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = Node(start[1] , start[0] , goal[1] , goal[0] , lowerCamelCase__ ) _lowerCamelCase = Node(goal[1] , goal[0] , goal[1] , goal[0] , lowerCamelCase__ ) _lowerCamelCase = [self.start] _lowerCamelCase = False def snake_case__ ( self ): while self.node_queue: _lowerCamelCase = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: _lowerCamelCase = True return self.retrace_path(lowerCamelCase__ ) _lowerCamelCase = self.get_successors(lowerCamelCase__ ) for node in successors: self.node_queue.append(lowerCamelCase__ ) if not self.reached: return [self.start.pos] return None def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for action in delta: _lowerCamelCase = parent.pos_x + action[1] _lowerCamelCase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCamelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(lowerCamelCase__ , lowerCamelCase__ , self.target.pos_y , self.target.pos_x , lowerCamelCase__ ) ) return successors def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = node _lowerCamelCase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _lowerCamelCase = current_node.parent path.reverse() return path class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BreadthFirstSearch(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = BreadthFirstSearch(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = False def snake_case__ ( self ): while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: _lowerCamelCase = self.fwd_bfs.node_queue.pop(0 ) _lowerCamelCase = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: _lowerCamelCase = True return self.retrace_bidirectional_path( lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = current_bwd_node _lowerCamelCase = current_fwd_node _lowerCamelCase = { self.fwd_bfs: self.fwd_bfs.get_successors(lowerCamelCase__ ), self.bwd_bfs: self.bwd_bfs.get_successors(lowerCamelCase__ ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(lowerCamelCase__ ) if not self.reached: return [self.fwd_bfs.start.pos] return None def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.fwd_bfs.retrace_path(lowerCamelCase__ ) _lowerCamelCase = self.bwd_bfs.retrace_path(lowerCamelCase__ ) bwd_path.pop() bwd_path.reverse() _lowerCamelCase = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : Dict = (0, 0) __SCREAMING_SNAKE_CASE : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __SCREAMING_SNAKE_CASE : List[Any] = time.time() __SCREAMING_SNAKE_CASE : List[str] = BreadthFirstSearch(init, goal) __SCREAMING_SNAKE_CASE : Optional[Any] = bfs.search() __SCREAMING_SNAKE_CASE : Dict = time.time() - start_bfs_time print('''Unidirectional BFS computation time : ''', bfs_time) __SCREAMING_SNAKE_CASE : Optional[int] = time.time() __SCREAMING_SNAKE_CASE : Any = BidirectionalBreadthFirstSearch(init, goal) __SCREAMING_SNAKE_CASE : int = bd_bfs.search() __SCREAMING_SNAKE_CASE : List[str] = time.time() - start_bd_bfs_time print('''Bidirectional BFS computation time : ''', bd_bfs_time)
73
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'mgp-str' def __init__( self , lowerCamelCase__=[3_2, 1_2_8] , lowerCamelCase__=4 , lowerCamelCase__=3 , lowerCamelCase__=2_7 , lowerCamelCase__=3_8 , lowerCamelCase__=5_0_2_5_7 , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=4.0 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=False , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = max_token_length _lowerCamelCase = num_character_labels _lowerCamelCase = num_bpe_labels _lowerCamelCase = num_wordpiece_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = mlp_ratio _lowerCamelCase = distilled _lowerCamelCase = layer_norm_eps _lowerCamelCase = drop_rate _lowerCamelCase = qkv_bias _lowerCamelCase = attn_drop_rate _lowerCamelCase = drop_path_rate _lowerCamelCase = output_aa_attentions _lowerCamelCase = initializer_range
73
1
import pprint import requests lowerCAmelCase = 'https://zenquotes.io/api' def _a ( ): """simple docstring""" return requests.get(API_ENDPOINT_URL + '''/today''' ).json() def _a ( ): """simple docstring""" return requests.get(API_ENDPOINT_URL + '''/random''' ).json() if __name__ == "__main__": lowerCAmelCase = random_quotes() pprint.pprint(response)
110
"""simple docstring""" from collections import defaultdict from math import ceil, sqrt def __SCREAMING_SNAKE_CASE ( A_ = 1_00_00_00 , A_ = 10 ): lowerCAmelCase__ : defaultdict = defaultdict(A_ ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: lowerCAmelCase__ : int = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: lowerCAmelCase__ : Tuple = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(A_ , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(F'''{solution() = }''')
106
0
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) __A : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name __A : Optional[Any] = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase=8 ) -> Any: '''simple docstring''' lowerCAmelCase : str = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCAmelCase : List[Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase=512, _UpperCAmelCase=512 ) -> int: '''simple docstring''' lowerCAmelCase : Tuple = pil_image.resize((w, h), resample=Image.BICUBIC, reducing_gap=1 ) lowerCAmelCase : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) lowerCAmelCase : List[str] = arr.astype(np.floataa ) / 1_2_7.5 - 1 lowerCAmelCase : List[Any] = np.transpose(_UpperCAmelCase, [2, 0, 1] ) lowerCAmelCase : Optional[int] = torch.from_numpy(_UpperCAmelCase ).unsqueeze(0 ) return image class __A ( lowerCAmelCase ): def __init__( self : List[Any] , UpperCAmelCase_ : UNetaDConditionModel , UpperCAmelCase_ : DDPMScheduler , UpperCAmelCase_ : VQModel , ): super().__init__() self.register_modules( unet=UpperCAmelCase_ , scheduler=UpperCAmelCase_ , movq=UpperCAmelCase_ , ) lowerCAmelCase : Optional[int] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowercase__ ( self : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Dict , UpperCAmelCase_ : int ): # get the original timestep using init_timestep lowerCAmelCase : int = min(int(num_inference_steps * strength ) , UpperCAmelCase_ ) lowerCAmelCase : int = max(num_inference_steps - init_timestep , 0 ) lowerCAmelCase : Optional[int] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowercase__ ( self : str , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[Any]=None ): if not isinstance(UpperCAmelCase_ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f"`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(UpperCAmelCase_ )}" ) lowerCAmelCase : str = image.to(device=UpperCAmelCase_ , dtype=UpperCAmelCase_ ) lowerCAmelCase : Optional[int] = batch_size * num_images_per_prompt if image.shape[1] == 4: lowerCAmelCase : List[str] = image else: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and len(UpperCAmelCase_ ) != batch_size: raise ValueError( f"You have passed a list of generators of length {len(UpperCAmelCase_ )}, but requested an effective batch" f" size of {batch_size}. Make sure the batch size matches the length of the generators." ) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase : Optional[int] = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(UpperCAmelCase_ ) ] lowerCAmelCase : Any = torch.cat(UpperCAmelCase_ , dim=0 ) else: lowerCAmelCase : Any = self.movq.encode(UpperCAmelCase_ ).latent_dist.sample(UpperCAmelCase_ ) lowerCAmelCase : int = self.movq.config.scaling_factor * init_latents lowerCAmelCase : Any = torch.cat([init_latents] , dim=0 ) lowerCAmelCase : Dict = init_latents.shape lowerCAmelCase : List[Any] = randn_tensor(UpperCAmelCase_ , generator=UpperCAmelCase_ , device=UpperCAmelCase_ , dtype=UpperCAmelCase_ ) # get latents lowerCAmelCase : int = self.scheduler.add_noise(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase : Optional[Any] = init_latents return latents def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Union[str, Any]=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowerCAmelCase : str = torch.device(f"cuda:{gpu_id}" ) lowerCAmelCase : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase_ , UpperCAmelCase_ ) def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : Union[str, Any]=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowerCAmelCase : List[str] = torch.device(f"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=UpperCAmelCase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCAmelCase : Dict = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCAmelCase , lowerCAmelCase : Optional[int] = cpu_offload_with_hook(UpperCAmelCase_ , UpperCAmelCase_ , prev_module_hook=UpperCAmelCase_ ) # We'll offload the last model manually. lowerCAmelCase : str = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowercase__ ( self : Any ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase_ ) def __call__( self : str , UpperCAmelCase_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase_ : Union[torch.FloatTensor, PIL.Image.Image, List[torch.FloatTensor], List[PIL.Image.Image]] , UpperCAmelCase_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase_ : int = 512 , UpperCAmelCase_ : int = 512 , UpperCAmelCase_ : int = 100 , UpperCAmelCase_ : float = 4.0 , UpperCAmelCase_ : float = 0.3 , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase_ : Optional[str] = "pil" , UpperCAmelCase_ : bool = True , ): lowerCAmelCase : Union[str, Any] = self._execution_device lowerCAmelCase : List[Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase : Optional[int] = torch.cat(UpperCAmelCase_ , dim=0 ) lowerCAmelCase : str = image_embeds.shape[0] if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase : Union[str, Any] = torch.cat(UpperCAmelCase_ , dim=0 ) if do_classifier_free_guidance: lowerCAmelCase : List[str] = image_embeds.repeat_interleave(UpperCAmelCase_ , dim=0 ) lowerCAmelCase : int = negative_image_embeds.repeat_interleave(UpperCAmelCase_ , dim=0 ) lowerCAmelCase : int = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase_ ) if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase : int = [image] if not all(isinstance(UpperCAmelCase_ , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( f"Input is in incorrect format: {[type(UpperCAmelCase_ ) for i in image]}. Currently, we only support PIL image and pytorch tensor" ) lowerCAmelCase : str = torch.cat([prepare_image(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) for i in image] , dim=0 ) lowerCAmelCase : Optional[Any] = image.to(dtype=image_embeds.dtype , device=UpperCAmelCase_ ) lowerCAmelCase : int = self.movq.encode(UpperCAmelCase_ )['latents'] lowerCAmelCase : Optional[int] = latents.repeat_interleave(UpperCAmelCase_ , dim=0 ) self.scheduler.set_timesteps(UpperCAmelCase_ , device=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase : List[str] = self.get_timesteps(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase : Tuple = timesteps[:1].repeat(batch_size * num_images_per_prompt ) lowerCAmelCase , lowerCAmelCase : Dict = downscale_height_and_width(UpperCAmelCase_ , UpperCAmelCase_ , self.movq_scale_factor ) lowerCAmelCase : Any = self.prepare_latents( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , image_embeds.dtype , UpperCAmelCase_ , UpperCAmelCase_ ) for i, t in enumerate(self.progress_bar(UpperCAmelCase_ ) ): # expand the latents if we are doing classifier free guidance lowerCAmelCase : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCAmelCase : List[str] = {'image_embeds': image_embeds} lowerCAmelCase : List[Any] = self.unet( sample=UpperCAmelCase_ , timestep=UpperCAmelCase_ , encoder_hidden_states=UpperCAmelCase_ , added_cond_kwargs=UpperCAmelCase_ , return_dict=UpperCAmelCase_ , )[0] if do_classifier_free_guidance: lowerCAmelCase , lowerCAmelCase : str = noise_pred.split(latents.shape[1] , dim=1 ) lowerCAmelCase , lowerCAmelCase : Any = noise_pred.chunk(2 ) lowerCAmelCase , lowerCAmelCase : int = variance_pred.chunk(2 ) lowerCAmelCase : Dict = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCAmelCase : List[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCAmelCase , lowerCAmelCase : Optional[int] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase : List[str] = self.scheduler.step( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , generator=UpperCAmelCase_ , )[0] # post-processing lowerCAmelCase : Optional[Any] = self.movq.decode(UpperCAmelCase_ , force_not_quantize=UpperCAmelCase_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: lowerCAmelCase : Union[str, Any] = image * 0.5 + 0.5 lowerCAmelCase : Union[str, Any] = image.clamp(0 , 1 ) lowerCAmelCase : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCAmelCase : Dict = self.numpy_to_pil(UpperCAmelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase_ )
323
from manim import * class __A ( lowerCAmelCase ): def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) lowerCAmelCase : Any = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) lowerCAmelCase : List[str] = Rectangle(height=0.25 , width=0.25 ) lowerCAmelCase : List[Any] = [mem.copy() for i in range(6 )] lowerCAmelCase : Tuple = [mem.copy() for i in range(6 )] lowerCAmelCase : int = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : Dict = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : int = VGroup(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : str = Text('CPU' , font_size=24 ) lowerCAmelCase : Union[str, Any] = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(UpperCAmelCase_ ) lowerCAmelCase : int = [mem.copy() for i in range(4 )] lowerCAmelCase : Union[str, Any] = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : int = Text('GPU' , font_size=24 ) lowerCAmelCase : Tuple = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) gpu.move_to([-1, -1, 0] ) self.add(UpperCAmelCase_ ) lowerCAmelCase : Union[str, Any] = [mem.copy() for i in range(6 )] lowerCAmelCase : Tuple = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : List[str] = Text('Model' , font_size=24 ) lowerCAmelCase : Union[str, Any] = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) model.move_to([3, -1.0, 0] ) self.add(UpperCAmelCase_ ) lowerCAmelCase : Any = [] lowerCAmelCase : Dict = [] for i, rect in enumerate(UpperCAmelCase_ ): lowerCAmelCase : Optional[Any] = fill.copy().set_fill(UpperCAmelCase_ , opacity=0.8 ) target.move_to(UpperCAmelCase_ ) model_arr.append(UpperCAmelCase_ ) lowerCAmelCase : Union[str, Any] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(UpperCAmelCase_ , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(UpperCAmelCase_ ) self.add(*UpperCAmelCase_ , *UpperCAmelCase_ ) lowerCAmelCase : Dict = [meta_mem.copy() for i in range(6 )] lowerCAmelCase : Union[str, Any] = [meta_mem.copy() for i in range(6 )] lowerCAmelCase : Tuple = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : int = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : Tuple = VGroup(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) lowerCAmelCase : Union[str, Any] = Text('Disk' , font_size=24 ) lowerCAmelCase : Optional[Any] = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) disk.move_to([-4, -1.25, 0] ) self.add(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase : Optional[int] = MarkupText( f"<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase : Dict = MarkupText( f"<span fgcolor='{BLUE}'>●</span> Checkpoint" , font_size=18 , ) blue_text.next_to(UpperCAmelCase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(UpperCAmelCase_ ) lowerCAmelCase : str = MarkupText( f"Now watch as an input is passed through the model\nand how the memory is utilized and handled." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCAmelCase_ ) ) lowerCAmelCase : Optional[Any] = Square(0.3 ) input.set_fill(UpperCAmelCase_ , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , UpperCAmelCase_ , buff=0.5 ) self.play(Write(UpperCAmelCase_ ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=UpperCAmelCase_ , buff=0.02 ) self.play(MoveToTarget(UpperCAmelCase_ ) ) self.play(FadeOut(UpperCAmelCase_ ) ) lowerCAmelCase : List[Any] = Arrow(start=UpperCAmelCase_ , end=UpperCAmelCase_ , color=UpperCAmelCase_ , buff=0.5 ) a.next_to(model_arr[0].get_left() , UpperCAmelCase_ , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) lowerCAmelCase : int = MarkupText( f"As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCAmelCase_ , run_time=3 ) ) lowerCAmelCase : Optional[Any] = {'run_time': 1, 'fade_in': True, 'fade_out': True, 'buff': 0.02} self.play( Write(UpperCAmelCase_ ) , Circumscribe(model_arr[0] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(model_cpu_arr[0] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) lowerCAmelCase : Any = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , UpperCAmelCase_ , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) lowerCAmelCase : int = AnimationGroup( FadeOut(UpperCAmelCase_ , run_time=0.5 ) , MoveToTarget(UpperCAmelCase_ , run_time=0.5 ) , FadeIn(UpperCAmelCase_ , run_time=0.5 ) , lag_ratio=0.2 ) self.play(UpperCAmelCase_ ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: lowerCAmelCase : List[str] = 0.7 self.play( Circumscribe(model_arr[i] , **UpperCAmelCase_ ) , Circumscribe(cpu_left_col_base[i] , **UpperCAmelCase_ ) , Circumscribe(cpu_left_col_base[i + 1] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(model_arr[i + 1] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(cpu_left_col_base[-1] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCAmelCase_ , **UpperCAmelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) lowerCAmelCase : int = a_c lowerCAmelCase : Union[str, Any] = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(UpperCAmelCase_ ) , FadeOut(UpperCAmelCase_ , run_time=0.5 ) , ) lowerCAmelCase : int = MarkupText(f"Inference on a model too large for GPU memory\nis successfully completed." , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCAmelCase_ , run_time=3 ) , MoveToTarget(UpperCAmelCase_ ) ) self.wait()
323
1
"""simple docstring""" from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer import PriorTransformer from .ta_film_transformer import TaFilmDecoder from .transformer_ad import TransformeraDModel from .unet_ad import UNetaDModel from .unet_ad import UNetaDModel from .unet_ad_condition import UNetaDConditionModel from .unet_ad_condition import UNetaDConditionModel from .vq_model import VQModel if is_flax_available(): from .controlnet_flax import FlaxControlNetModel from .unet_ad_condition_flax import FlaxUNetaDConditionModel from .vae_flax import FlaxAutoencoderKL
108
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCamelCase : """simple docstring""" def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=[10, 20, 30, 40] , _SCREAMING_SNAKE_CASE=[1, 1, 2, 1] , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE="relu" , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=None , )->List[str]: '''simple docstring''' A_ : str = parent A_ : int = batch_size A_ : List[str] = image_size A_ : Dict = num_channels A_ : Tuple = embeddings_size A_ : Union[str, Any] = hidden_sizes A_ : Dict = depths A_ : str = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_act A_ : Optional[Any] = num_labels A_ : Tuple = scope A_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Optional[Any]: '''simple docstring''' A_ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : str = None if self.use_labels: A_ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def _snake_case ( self )->Union[str, Any]: '''simple docstring''' return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' A_ : Dict = RegNetModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ : Any = model(_SCREAMING_SNAKE_CASE ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' A_ : Union[str, Any] = self.num_labels A_ : Dict = RegNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ : int = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self )->Union[str, Any]: '''simple docstring''' A_ : Tuple = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Any = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _lowerCamelCase ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" snake_case = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () snake_case = ( {"feature-extraction": RegNetModel, "image-classification": RegNetForImageClassification} if is_torch_available() else {} ) snake_case = False snake_case = False snake_case = False snake_case = False def _snake_case ( self )->Union[str, Any]: '''simple docstring''' A_ : Union[str, Any] = RegNetModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Dict: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _snake_case ( self )->Tuple: '''simple docstring''' return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def _snake_case ( self )->Dict: '''simple docstring''' pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def _snake_case ( self )->str: '''simple docstring''' pass def _snake_case ( self )->List[Any]: '''simple docstring''' A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : str = model_class(_SCREAMING_SNAKE_CASE ) A_ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Any = [*signature.parameters.keys()] A_ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Any: '''simple docstring''' A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Optional[Any]: '''simple docstring''' A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Union[str, Any] = model_class(config=_SCREAMING_SNAKE_CASE ) for name, module in model.named_modules(): if isinstance(_SCREAMING_SNAKE_CASE , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) def _snake_case ( self )->List[Any]: '''simple docstring''' def check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): A_ : str = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): A_ : Tuple = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) A_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() A_ : int = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : int = layer_type A_ : List[Any] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Dict: '''simple docstring''' A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def _snake_case ( self )->str: '''simple docstring''' for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = RegNetModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def _SCREAMING_SNAKE_CASE ( ): A_ : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _lowerCamelCase ( unittest.TestCase ): """simple docstring""" @cached_property def _snake_case ( self )->List[str]: '''simple docstring''' return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _snake_case ( self )->Tuple: '''simple docstring''' A_ : List[Any] = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_SCREAMING_SNAKE_CASE ) A_ : Optional[Any] = self.default_image_processor A_ : Any = prepare_img() A_ : Optional[Any] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A_ : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) # verify the logits A_ : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) A_ : Optional[int] = torch.tensor([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) )
186
0
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : int = {"""configuration_timm_backbone""": ["""TimmBackboneConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Tuple = ["""TimmBackbone"""] if TYPE_CHECKING: from .configuration_timm_backbone import TimmBackboneConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timm_backbone import TimmBackbone else: import sys _UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
353
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowercase ( lowercase_ , unittest.TestCase ): __SCREAMING_SNAKE_CASE : List[str] = BarthezTokenizer __SCREAMING_SNAKE_CASE : str = BarthezTokenizerFast __SCREAMING_SNAKE_CASE : Optional[Any] = True __SCREAMING_SNAKE_CASE : str = True def a ( self ): super().setUp() snake_case_ = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case ) snake_case_ = tokenizer def a ( self ): snake_case_ = '<pad>' snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case ) , snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case ) , snake_case ) def a ( self ): snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(snake_case ) , 10_1122 ) def a ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_1122 ) @require_torch def a ( self ): snake_case_ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] snake_case_ = [0, 57, 3018, 7_0307, 91, 2] snake_case_ = self.tokenizer( snake_case , max_length=len(snake_case ) , padding=snake_case , truncation=snake_case , return_tensors='pt' ) self.assertIsInstance(snake_case , snake_case ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) snake_case_ = batch.input_ids.tolist()[0] self.assertListEqual(snake_case , snake_case ) def a ( self ): if not self.test_rust_tokenizer: return snake_case_ = self.get_tokenizer() snake_case_ = self.get_rust_tokenizer() snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = tokenizer.tokenize(snake_case ) snake_case_ = rust_tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , snake_case ) snake_case_ = tokenizer.encode(snake_case , add_special_tokens=snake_case ) snake_case_ = rust_tokenizer.encode(snake_case , add_special_tokens=snake_case ) self.assertListEqual(snake_case , snake_case ) snake_case_ = self.get_rust_tokenizer() snake_case_ = tokenizer.encode(snake_case ) snake_case_ = rust_tokenizer.encode(snake_case ) self.assertListEqual(snake_case , snake_case ) @slow def a ( self ): # fmt: off snake_case_ = {'input_ids': [[0, 490, 1_4328, 4507, 354, 47, 4_3669, 95, 25, 7_8117, 2_0215, 1_9779, 190, 22, 400, 4, 3_5343, 8_0310, 603, 86, 2_4937, 105, 3_3438, 9_4762, 196, 3_9642, 7, 15, 1_5933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0534, 87, 25, 66, 3358, 196, 5_5289, 8, 8_2961, 81, 2204, 7_5203, 7, 15, 763, 1_2956, 216, 178, 1_4328, 9595, 1377, 6_9693, 7, 448, 7_1021, 196, 1_8106, 1437, 1_3974, 108, 9083, 4, 4_9315, 7, 39, 86, 1326, 2793, 4_6333, 4, 448, 196, 7_4588, 7, 4_9315, 7, 39, 21, 822, 3_8470, 74, 21, 6_6723, 6_2480, 8, 2_2050, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. snake_case_ = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=snake_case , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=snake_case , )
200
0
'''simple docstring''' import sys SCREAMING_SNAKE_CASE_: Optional[int] =( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def lowerCAmelCase_ ( snake_case_ : str ) -> int: '''simple docstring''' UpperCAmelCase_ = 1 for digit in s: product *= int(snake_case_ ) return product def lowerCAmelCase_ ( snake_case_ : str = N ) -> int: '''simple docstring''' UpperCAmelCase_ = -sys.maxsize - 1 UpperCAmelCase_ = n[:13] UpperCAmelCase_ = 13 while cur_index < len(snake_case_ ) - 13: if int(n[cur_index] ) >= int(substr[0] ): UpperCAmelCase_ = substr[1:] + n[cur_index] cur_index += 1 else: UpperCAmelCase_ = max(snake_case_ , str_eval(snake_case_ ) ) UpperCAmelCase_ = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f"{solution() = }")
1
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE_: Optional[int] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE_: Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append( ( f"transformer.decoder.layers.{i}.cross_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", ) ) rename_keys.append( ( f"transformer.decoder.layers.{i}.cross_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", ) ) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f"transformer.decoder.layers.{i}.sa_qcontent_proj.weight", f"decoder.layers.{i}.sa_qcontent_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.sa_kcontent_proj.weight", f"decoder.layers.{i}.sa_kcontent_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.sa_qpos_proj.weight", f"decoder.layers.{i}.sa_qpos_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.sa_kpos_proj.weight", f"decoder.layers.{i}.sa_kpos_proj.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.sa_v_proj.weight", f"decoder.layers.{i}.sa_v_proj.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_qcontent_proj.weight", f"decoder.layers.{i}.ca_qcontent_proj.weight") ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_kcontent_proj.weight", f"decoder.layers.{i}.ca_kcontent_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_kpos_proj.weight", f"decoder.layers.{i}.ca_kpos_proj.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.ca_v_proj.weight", f"decoder.layers.{i}.ca_v_proj.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight", f"decoder.layers.{i}.ca_qpos_sine_proj.weight") ) rename_keys.append( (f"transformer.decoder.layers.{i}.sa_qcontent_proj.bias", f"decoder.layers.{i}.sa_qcontent_proj.bias") ) rename_keys.append( (f"transformer.decoder.layers.{i}.sa_kcontent_proj.bias", f"decoder.layers.{i}.sa_kcontent_proj.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.sa_qpos_proj.bias", f"decoder.layers.{i}.sa_qpos_proj.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.sa_kpos_proj.bias", f"decoder.layers.{i}.sa_kpos_proj.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.sa_v_proj.bias", f"decoder.layers.{i}.sa_v_proj.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_qcontent_proj.bias", f"decoder.layers.{i}.ca_qcontent_proj.bias") ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_kcontent_proj.bias", f"decoder.layers.{i}.ca_kcontent_proj.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.ca_kpos_proj.bias", f"decoder.layers.{i}.ca_kpos_proj.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.ca_v_proj.bias", f"decoder.layers.{i}.ca_v_proj.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias", f"decoder.layers.{i}.ca_qpos_sine_proj.bias") ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def lowerCAmelCase_ ( snake_case_ : Dict , snake_case_ : Any , snake_case_ : Optional[int] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = state_dict.pop(snake_case_ ) UpperCAmelCase_ = val def lowerCAmelCase_ ( snake_case_ : int ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: UpperCAmelCase_ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) UpperCAmelCase_ = value else: UpperCAmelCase_ = value return new_state_dict def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Dict=False ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = "" if is_panoptic: UpperCAmelCase_ = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:2_56, :] UpperCAmelCase_ = in_proj_bias[:2_56] UpperCAmelCase_ = in_proj_weight[2_56:5_12, :] UpperCAmelCase_ = in_proj_bias[2_56:5_12] UpperCAmelCase_ = in_proj_weight[-2_56:, :] UpperCAmelCase_ = in_proj_bias[-2_56:] def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(snake_case_ , stream=snake_case_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( snake_case_ : Optional[int] , snake_case_ : Dict ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: UpperCAmelCase_ = "resnet101" if "dc5" in model_name: UpperCAmelCase_ = True UpperCAmelCase_ = "panoptic" in model_name if is_panoptic: UpperCAmelCase_ = 2_50 else: UpperCAmelCase_ = 91 UpperCAmelCase_ = "huggingface/label-files" UpperCAmelCase_ = "coco-detection-id2label.json" UpperCAmelCase_ = json.load(open(hf_hub_download(snake_case_ , snake_case_ , repo_type="dataset" ) , "r" ) ) UpperCAmelCase_ = {int(snake_case_ ): v for k, v in idalabel.items()} UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} # load image processor UpperCAmelCase_ = "coco_panoptic" if is_panoptic else "coco_detection" UpperCAmelCase_ = ConditionalDetrImageProcessor(format=snake_case_ ) # prepare image UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=snake_case_ , return_tensors="pt" ) UpperCAmelCase_ = encoding["pixel_values"] logger.info(f"""Converting model {model_name}...""" ) # load original model from torch hub UpperCAmelCase_ = torch.hub.load("DeppMeng/ConditionalDETR" , snake_case_ , pretrained=snake_case_ ).eval() UpperCAmelCase_ = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: UpperCAmelCase_ = "conditional_detr." + src rename_key(snake_case_ , snake_case_ , snake_case_ ) UpperCAmelCase_ = rename_backbone_keys(snake_case_ ) # query, key and value matrices need special treatment read_in_q_k_v(snake_case_ , is_panoptic=snake_case_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCAmelCase_ = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): UpperCAmelCase_ = state_dict.pop(snake_case_ ) UpperCAmelCase_ = val elif "class_labels_classifier" in key or "bbox_predictor" in key: UpperCAmelCase_ = state_dict.pop(snake_case_ ) UpperCAmelCase_ = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: UpperCAmelCase_ = state_dict.pop(snake_case_ ) UpperCAmelCase_ = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): UpperCAmelCase_ = state_dict.pop(snake_case_ ) UpperCAmelCase_ = val # finally, create HuggingFace model and load state dict UpperCAmelCase_ = ConditionalDetrForSegmentation(snake_case_ ) if is_panoptic else ConditionalDetrForObjectDetection(snake_case_ ) model.load_state_dict(snake_case_ ) model.eval() model.push_to_hub(repo_id=snake_case_ , organization="DepuMeng" , commit_message="Add model" ) # verify our conversion UpperCAmelCase_ = conditional_detr(snake_case_ ) UpperCAmelCase_ = model(snake_case_ ) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4 ) # Save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(snake_case_ ).mkdir(exist_ok=snake_case_ ) model.save_pretrained(snake_case_ ) image_processor.save_pretrained(snake_case_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: List[str] =argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) SCREAMING_SNAKE_CASE_: int =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
1
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __magic_name__ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase): A: Tuple = CycleDiffusionPipeline A: int = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "negative_prompt", "height", "width", "negative_prompt_embeds", } A: str = PipelineTesterMixin.required_optional_params - {"latents"} A: Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"source_prompt"}) A: List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS A: Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase__ ( self : List[Any] ) -> int: '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase__ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCamelCase__ : List[str] = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , num_train_timesteps=1000 , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) torch.manual_seed(0 ) UpperCamelCase__ : Dict = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCamelCase__ : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) UpperCamelCase__ : Tuple = CLIPTextModel(lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) UpperCamelCase__ : Dict = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[int]=0 ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Dict = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) UpperCamelCase__ : Dict = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): UpperCamelCase__ : Optional[Any] = torch.manual_seed(lowerCamelCase__ ) else: UpperCamelCase__ : int = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCamelCase__ : Dict = { '''prompt''': '''An astronaut riding an elephant''', '''source_prompt''': '''An astronaut riding a horse''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''eta''': 0.1, '''strength''': 0.8, '''guidance_scale''': 3, '''source_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : str = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase__ : int = self.get_dummy_components() UpperCamelCase__ : Dict = CycleDiffusionPipeline(**lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = self.get_dummy_inputs(lowerCamelCase__ ) UpperCamelCase__ : List[str] = pipe(**lowerCamelCase__ ) UpperCamelCase__ : Dict = output.images UpperCamelCase__ : int = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) UpperCamelCase__ : Optional[int] = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : List[str] = self.get_dummy_components() for name, module in components.items(): if hasattr(lowerCamelCase__ , '''half''' ): UpperCamelCase__ : List[str] = module.half() UpperCamelCase__ : Optional[int] = CycleDiffusionPipeline(**lowerCamelCase__ ) UpperCamelCase__ : Any = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCamelCase__ : Optional[Any] = self.get_dummy_inputs(lowerCamelCase__ ) UpperCamelCase__ : int = pipe(**lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = output.images UpperCamelCase__ : Optional[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) UpperCamelCase__ : Dict = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def UpperCAmelCase__ ( self : int ) -> Union[str, Any]: '''simple docstring''' return super().test_save_load_local() @unittest.skip('''non-deterministic pipeline''' ) def UpperCAmelCase__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' return super().test_inference_batch_single_identical() @skip_mps def UpperCAmelCase__ ( self : Optional[int] ) -> int: '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def UpperCAmelCase__ ( self : str ) -> Union[str, Any]: '''simple docstring''' return super().test_save_load_optional_components() @skip_mps def UpperCAmelCase__ ( self : Optional[int] ) -> Any: '''simple docstring''' return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' UpperCamelCase__ : str = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) UpperCamelCase__ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy''' ) UpperCamelCase__ : List[Any] = init_image.resize((512, 512) ) UpperCamelCase__ : Tuple = '''CompVis/stable-diffusion-v1-4''' UpperCamelCase__ : Dict = DDIMScheduler.from_pretrained(lowerCamelCase__ , subfolder='''scheduler''' ) UpperCamelCase__ : Optional[int] = CycleDiffusionPipeline.from_pretrained( lowerCamelCase__ , scheduler=lowerCamelCase__ , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa , revision='''fp16''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() UpperCamelCase__ : Union[str, Any] = '''A black colored car''' UpperCamelCase__ : Any = '''A blue colored car''' UpperCamelCase__ : Union[str, Any] = torch.manual_seed(0 ) UpperCamelCase__ : Optional[int] = pipe( prompt=lowerCamelCase__ , source_prompt=lowerCamelCase__ , image=lowerCamelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowerCamelCase__ , output_type='''np''' , ) UpperCamelCase__ : Any = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def UpperCAmelCase__ ( self : Any ) -> List[str]: '''simple docstring''' UpperCamelCase__ : Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) UpperCamelCase__ : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy''' ) UpperCamelCase__ : Tuple = init_image.resize((512, 512) ) UpperCamelCase__ : str = '''CompVis/stable-diffusion-v1-4''' UpperCamelCase__ : Optional[Any] = DDIMScheduler.from_pretrained(lowerCamelCase__ , subfolder='''scheduler''' ) UpperCamelCase__ : Union[str, Any] = CycleDiffusionPipeline.from_pretrained(lowerCamelCase__ , scheduler=lowerCamelCase__ , safety_checker=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() UpperCamelCase__ : Tuple = '''A black colored car''' UpperCamelCase__ : List[Any] = '''A blue colored car''' UpperCamelCase__ : Union[str, Any] = torch.manual_seed(0 ) UpperCamelCase__ : List[Any] = pipe( prompt=lowerCamelCase__ , source_prompt=lowerCamelCase__ , image=lowerCamelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowerCamelCase__ , output_type='''np''' , ) UpperCamelCase__ : Dict = output.images assert np.abs(image - expected_image ).max() < 2E-2
352
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def _a ( SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" UpperCamelCase__ , UpperCamelCase__ : Dict = image.size UpperCamelCase__ , UpperCamelCase__ : List[Any] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 UpperCamelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) UpperCamelCase__ : Union[str, Any] = np.array(SCREAMING_SNAKE_CASE ).astype(np.floataa ) / 255.0 UpperCamelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) UpperCamelCase__ : int = torch.from_numpy(SCREAMING_SNAKE_CASE ) return 2.0 * image - 1.0 class __magic_name__ ( __lowerCAmelCase): def __init__( self : Dict , lowerCamelCase__ : VQModel , lowerCamelCase__ : UNetaDModel , lowerCamelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Tuple: '''simple docstring''' super().__init__() self.register_modules(vqvae=lowerCamelCase__ , unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) @torch.no_grad() def __call__( self : int , lowerCamelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase__ : Optional[int] = 1 , lowerCamelCase__ : Optional[int] = 100 , lowerCamelCase__ : Optional[float] = 0.0 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' if isinstance(lowerCamelCase__ , PIL.Image.Image ): UpperCamelCase__ : int = 1 elif isinstance(lowerCamelCase__ , torch.Tensor ): UpperCamelCase__ : Dict = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase__ )}" ) if isinstance(lowerCamelCase__ , PIL.Image.Image ): UpperCamelCase__ : Any = preprocess(lowerCamelCase__ ) UpperCamelCase__ , UpperCamelCase__ : Tuple = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image UpperCamelCase__ : Any = (batch_size, self.unet.config.in_channels // 2, height, width) UpperCamelCase__ : Union[str, Any] = next(self.unet.parameters() ).dtype UpperCamelCase__ : Any = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=self.device , dtype=lowerCamelCase__ ) UpperCamelCase__ : Any = image.to(device=self.device , dtype=lowerCamelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) UpperCamelCase__ : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler UpperCamelCase__ : int = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCamelCase__ : Dict = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCamelCase__ : Optional[int] = {} if accepts_eta: UpperCamelCase__ : Union[str, Any] = eta for t in self.progress_bar(lowerCamelCase__ ): # concat latents and low resolution image in the channel dimension. UpperCamelCase__ : Any = torch.cat([latents, image] , dim=1 ) UpperCamelCase__ : List[str] = self.scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) # predict the noise residual UpperCamelCase__ : Dict = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase__ : Tuple = self.scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample # decode the image latents with the VQVAE UpperCamelCase__ : Tuple = self.vqvae.decode(lowerCamelCase__ ).sample UpperCamelCase__ : Tuple = torch.clamp(lowerCamelCase__ , -1.0 , 1.0 ) UpperCamelCase__ : Any = image / 2 + 0.5 UpperCamelCase__ : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase__ : List[str] = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase__ )
51
0