code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=__UpperCAmelCase ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) snake_case__ = Features({"text": Value("string" )} ) snake_case__ = Features({"labels": ClassLabel} ) snake_case__ = "text" snake_case__ = "labels" def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Optional[Any] ): if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] ,lowerCamelCase__ ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) UpperCAmelCase__ = copy.deepcopy(self ) UpperCAmelCase__ = self.label_schema.copy() UpperCAmelCase__ = features[self.label_column] UpperCAmelCase__ = label_schema return task_template @property def __lowerCAmelCase ( self : List[str] ): return { self.text_column: "text", self.label_column: "labels", }
702
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
632
0
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = XLMProphetNetTokenizer snake_case__ = False snake_case__ = True def __lowerCAmelCase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = '[PAD]' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'[PAD]' ) self.assertEqual(vocab_keys[1] ,'[CLS]' ) self.assertEqual(vocab_keys[-1] ,'j' ) self.assertEqual(len(lowerCamelCase__ ) ,1_012 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_012 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) UpperCAmelCase__ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] ,) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] ,) @cached_property def __lowerCAmelCase ( self : Dict ): return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'Hello World!' UpperCAmelCase__ = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/xprophetnet-large-wiki100-cased' ,revision='1acad1643ddd54a44df6a1b797ada8373685d90e' ,)
703
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
632
0
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCAmelCase__ : Optional[int] = 'pt' elif is_tf_available(): lowerCAmelCase__ : Union[str, Any] = 'tf' else: lowerCAmelCase__ : Tuple = 'jax' class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = PerceiverTokenizer snake_case__ = False def __lowerCAmelCase ( self : str ): super().setUp() UpperCAmelCase__ = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def __lowerCAmelCase ( self : Union[str, Any] ): return PerceiverTokenizer.from_pretrained('deepmind/language-perceiver' ) def __lowerCAmelCase ( self : Union[str, Any] ,**lowerCamelCase__ : Optional[int] ): return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Optional[Any]=20 ,lowerCamelCase__ : str=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): try: UpperCAmelCase__ = tokenizer.decode([i] ,clean_up_tokenization_spaces=lowerCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) UpperCAmelCase__ = list(filter(lambda lowerCamelCase__ : re.match(R'^[ a-zA-Z]+$' ,t[1] ) ,lowerCamelCase__ ) ) UpperCAmelCase__ = list(filter(lambda lowerCamelCase__ : [t[0]] == tokenizer.encode(t[1] ,add_special_tokens=lowerCamelCase__ ) ,lowerCamelCase__ ) ) if max_length is not None and len(lowerCamelCase__ ) > max_length: UpperCAmelCase__ = toks[:max_length] if min_length is not None and len(lowerCamelCase__ ) < min_length and len(lowerCamelCase__ ) > 0: while len(lowerCamelCase__ ) < min_length: UpperCAmelCase__ = toks + toks # toks_str = [t[1] for t in toks] UpperCAmelCase__ = [t[0] for t in toks] # Ensure consistency UpperCAmelCase__ = tokenizer.decode(lowerCamelCase__ ,clean_up_tokenization_spaces=lowerCamelCase__ ) if " " not in output_txt and len(lowerCamelCase__ ) > 1: UpperCAmelCase__ = ( tokenizer.decode([toks_ids[0]] ,clean_up_tokenization_spaces=lowerCamelCase__ ) + ' ' + tokenizer.decode(toks_ids[1:] ,clean_up_tokenization_spaces=lowerCamelCase__ ) ) if with_prefix_space: UpperCAmelCase__ = ' ' + output_txt UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) return output_txt, output_ids def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.perceiver_tokenizer UpperCAmelCase__ = 'Unicode €.' UpperCAmelCase__ = tokenizer(lowerCamelCase__ ) UpperCAmelCase__ = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded['input_ids'] ,lowerCamelCase__ ) # decoding UpperCAmelCase__ = tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,'[CLS]Unicode €.[SEP]' ) UpperCAmelCase__ = tokenizer('e è é ê ë' ) UpperCAmelCase__ = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded['input_ids'] ,lowerCamelCase__ ) # decoding UpperCAmelCase__ = tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,'[CLS]e è é ê ë[SEP]' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ) ,'[CLS]e è é ê ë[SEP]' ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.perceiver_tokenizer UpperCAmelCase__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off UpperCAmelCase__ = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on UpperCAmelCase__ = tokenizer(lowerCamelCase__ ,padding=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ ,lowerCamelCase__ ) if FRAMEWORK != "jax": UpperCAmelCase__ = list(batch.input_ids.numpy()[0] ) else: UpperCAmelCase__ = list(batch.input_ids.tolist()[0] ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual((2, 38) ,batch.input_ids.shape ) self.assertEqual((2, 38) ,batch.attention_mask.shape ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = self.perceiver_tokenizer UpperCAmelCase__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] UpperCAmelCase__ = tokenizer(lowerCamelCase__ ,padding=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids' ,lowerCamelCase__ ) self.assertIn('attention_mask' ,lowerCamelCase__ ) self.assertNotIn('decoder_input_ids' ,lowerCamelCase__ ) self.assertNotIn('decoder_attention_mask' ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.perceiver_tokenizer UpperCAmelCase__ = [ 'Summary of the text.', 'Another summary.', ] UpperCAmelCase__ = tokenizer( text_target=lowerCamelCase__ ,max_length=32 ,padding='max_length' ,truncation=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ) self.assertEqual(32 ,targets['input_ids'].shape[1] ) def __lowerCAmelCase ( self : List[Any] ): # safety check on max_len default value so we are sure the test works UpperCAmelCase__ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length ,42 ) # Now let's start the test UpperCAmelCase__ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = ' He is very happy, UNwant\u00E9d,running' UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) tokenizer.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.__class__.from_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = after_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) shutil.rmtree(lowerCamelCase__ ) UpperCAmelCase__ = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) UpperCAmelCase__ = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) tokenizer.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.__class__.from_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = after_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) self.assertIn('new_additional_special_token' ,after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length ,42 ) UpperCAmelCase__ = tokenizer.__class__.from_pretrained(lowerCamelCase__ ,model_max_length=43 ) self.assertEqual(tokenizer.model_max_length ,43 ) shutil.rmtree(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ ,'special_tokens_map.json' ) ,encoding='utf-8' ) as json_file: UpperCAmelCase__ = json.load(lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ ,'tokenizer_config.json' ) ,encoding='utf-8' ) as json_file: UpperCAmelCase__ = json.load(lowerCamelCase__ ) UpperCAmelCase__ = [f'''<extra_id_{i}>''' for i in range(125 )] UpperCAmelCase__ = added_tokens_extra_ids + [ 'an_additional_special_token' ] UpperCAmelCase__ = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(lowerCamelCase__ ,'special_tokens_map.json' ) ,'w' ,encoding='utf-8' ) as outfile: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ ,'tokenizer_config.json' ) ,'w' ,encoding='utf-8' ) as outfile: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files UpperCAmelCase__ = tokenizer_class.from_pretrained( lowerCamelCase__ ,) self.assertIn( 'an_additional_special_token' ,tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['an_additional_special_token'] ,tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ) ,) # Now we test that we can change the value of additional_special_tokens in the from_pretrained UpperCAmelCase__ = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token' ,lstrip=lowerCamelCase__ )] UpperCAmelCase__ = tokenizer_class.from_pretrained( lowerCamelCase__ ,additional_special_tokens=lowerCamelCase__ ,) self.assertIn('a_new_additional_special_token' ,tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'] ,tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ) ,) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) ,'�' ) def __lowerCAmelCase ( self : List[str] ): pass def __lowerCAmelCase ( self : Optional[int] ): pass def __lowerCAmelCase ( self : Optional[int] ): pass def __lowerCAmelCase ( self : Tuple ): pass def __lowerCAmelCase ( self : int ): # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens UpperCAmelCase__ = self.get_tokenizers(fast=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): UpperCAmelCase__ = ['[CLS]', 't', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 's', 't', '[SEP]'] UpperCAmelCase__ = tokenizer.convert_tokens_to_string(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ ,lowerCamelCase__ )
704
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
632
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowerCAmelCase__ : int = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP', 'SpeechT5Config', 'SpeechT5HifiGanConfig', ], 'feature_extraction_speecht5': ['SpeechT5FeatureExtractor'], 'processing_speecht5': ['SpeechT5Processor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[Any] = ['SpeechT5Tokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'SpeechT5ForSpeechToText', 'SpeechT5ForSpeechToSpeech', 'SpeechT5ForTextToSpeech', 'SpeechT5Model', 'SpeechT5PreTrainedModel', 'SpeechT5HifiGan', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowerCAmelCase__ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
705
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
632
0
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = 42 class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @register_to_config def __init__( self : Any ,lowerCamelCase__ : int = 32 ,lowerCamelCase__ : int = 64 ,lowerCamelCase__ : int = 20 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : Optional[Any]=77 ,lowerCamelCase__ : Optional[int]=4 ,lowerCamelCase__ : float = 0.0 ,lowerCamelCase__ : str = "silu" ,lowerCamelCase__ : Optional[str] = None ,lowerCamelCase__ : Optional[str] = None ,lowerCamelCase__ : Optional[str] = "linear" ,lowerCamelCase__ : Optional[str] = "prd" ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,): super().__init__() UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = attention_head_dim UpperCAmelCase__ = num_attention_heads * attention_head_dim UpperCAmelCase__ = additional_embeddings UpperCAmelCase__ = time_embed_dim or inner_dim UpperCAmelCase__ = embedding_proj_dim or embedding_dim UpperCAmelCase__ = clip_embed_dim or embedding_dim UpperCAmelCase__ = Timesteps(lowerCamelCase__ ,lowerCamelCase__ ,0 ) UpperCAmelCase__ = TimestepEmbedding(lowerCamelCase__ ,lowerCamelCase__ ,out_dim=lowerCamelCase__ ,act_fn=lowerCamelCase__ ) UpperCAmelCase__ = nn.Linear(lowerCamelCase__ ,lowerCamelCase__ ) if embedding_proj_norm_type is None: UpperCAmelCase__ = None elif embedding_proj_norm_type == "layer": UpperCAmelCase__ = nn.LayerNorm(lowerCamelCase__ ) else: raise ValueError(f'''unsupported embedding_proj_norm_type: {embedding_proj_norm_type}''' ) UpperCAmelCase__ = nn.Linear(lowerCamelCase__ ,lowerCamelCase__ ) if encoder_hid_proj_type is None: UpperCAmelCase__ = None elif encoder_hid_proj_type == "linear": UpperCAmelCase__ = nn.Linear(lowerCamelCase__ ,lowerCamelCase__ ) else: raise ValueError(f'''unsupported encoder_hid_proj_type: {encoder_hid_proj_type}''' ) UpperCAmelCase__ = nn.Parameter(torch.zeros(1 ,num_embeddings + additional_embeddings ,lowerCamelCase__ ) ) if added_emb_type == "prd": UpperCAmelCase__ = nn.Parameter(torch.zeros(1 ,1 ,lowerCamelCase__ ) ) elif added_emb_type is None: UpperCAmelCase__ = None else: raise ValueError( f'''`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `\'prd\'` or `None`.''' ) UpperCAmelCase__ = nn.ModuleList( [ BasicTransformerBlock( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,dropout=lowerCamelCase__ ,activation_fn='gelu' ,attention_bias=lowerCamelCase__ ,) for d in range(lowerCamelCase__ ) ] ) if norm_in_type == "layer": UpperCAmelCase__ = nn.LayerNorm(lowerCamelCase__ ) elif norm_in_type is None: UpperCAmelCase__ = None else: raise ValueError(f'''Unsupported norm_in_type: {norm_in_type}.''' ) UpperCAmelCase__ = nn.LayerNorm(lowerCamelCase__ ) UpperCAmelCase__ = nn.Linear(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] ,-10_000.0 ) causal_attention_mask.triu_(1 ) UpperCAmelCase__ = causal_attention_mask[None, ...] self.register_buffer('causal_attention_mask' ,lowerCamelCase__ ,persistent=lowerCamelCase__ ) UpperCAmelCase__ = nn.Parameter(torch.zeros(1 ,lowerCamelCase__ ) ) UpperCAmelCase__ = nn.Parameter(torch.zeros(1 ,lowerCamelCase__ ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = {} def fn_recursive_add_processors(lowerCamelCase__ : str ,lowerCamelCase__ : torch.nn.Module ,lowerCamelCase__ : Dict[str, AttentionProcessor] ): if hasattr(lowerCamelCase__ ,'set_processor' ): UpperCAmelCase__ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' ,lowerCamelCase__ ,lowerCamelCase__ ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return processors def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ): UpperCAmelCase__ = len(self.attn_processors.keys() ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and len(lowerCamelCase__ ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(lowerCamelCase__ )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(lowerCamelCase__ : str ,lowerCamelCase__ : torch.nn.Module ,lowerCamelCase__ : List[str] ): if hasattr(lowerCamelCase__ ,'set_processor' ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): module.set_processor(lowerCamelCase__ ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' ,lowerCamelCase__ ,lowerCamelCase__ ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : int ): self.set_attn_processor(AttnProcessor() ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Union[torch.Tensor, float, int] ,lowerCamelCase__ : torch.FloatTensor ,lowerCamelCase__ : Optional[torch.FloatTensor] = None ,lowerCamelCase__ : Optional[torch.BoolTensor] = None ,lowerCamelCase__ : bool = True ,): UpperCAmelCase__ = hidden_states.shape[0] UpperCAmelCase__ = timestep if not torch.is_tensor(lowerCamelCase__ ): UpperCAmelCase__ = torch.tensor([timesteps] ,dtype=torch.long ,device=hidden_states.device ) elif torch.is_tensor(lowerCamelCase__ ) and len(timesteps.shape ) == 0: UpperCAmelCase__ = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML UpperCAmelCase__ = timesteps * torch.ones(lowerCamelCase__ ,dtype=timesteps.dtype ,device=timesteps.device ) UpperCAmelCase__ = self.time_proj(lowerCamelCase__ ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. UpperCAmelCase__ = timesteps_projected.to(dtype=self.dtype ) UpperCAmelCase__ = self.time_embedding(lowerCamelCase__ ) if self.embedding_proj_norm is not None: UpperCAmelCase__ = self.embedding_proj_norm(lowerCamelCase__ ) UpperCAmelCase__ = self.embedding_proj(lowerCamelCase__ ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: UpperCAmelCase__ = self.encoder_hidden_states_proj(lowerCamelCase__ ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError('`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set' ) UpperCAmelCase__ = self.proj_in(lowerCamelCase__ ) UpperCAmelCase__ = self.positional_embedding.to(hidden_states.dtype ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 if encoder_hidden_states is not None: additional_embeds.append(lowerCamelCase__ ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: UpperCAmelCase__ = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: UpperCAmelCase__ = hidden_states[:, None, :] UpperCAmelCase__ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: UpperCAmelCase__ = self.prd_embedding.to(hidden_states.dtype ).expand(lowerCamelCase__ ,-1 ,-1 ) additional_embeds.append(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat( lowerCamelCase__ ,dim=1 ,) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens UpperCAmelCase__ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: UpperCAmelCase__ = F.pad( lowerCamelCase__ ,( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) ,value=0.0 ,) UpperCAmelCase__ = hidden_states + positional_embeddings if attention_mask is not None: UpperCAmelCase__ = (1 - attention_mask.to(hidden_states.dtype )) * -10_000.0 UpperCAmelCase__ = F.pad(lowerCamelCase__ ,(0, self.additional_embeddings) ,value=0.0 ) UpperCAmelCase__ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) UpperCAmelCase__ = attention_mask.repeat_interleave(self.config.num_attention_heads ,dim=0 ) if self.norm_in is not None: UpperCAmelCase__ = self.norm_in(lowerCamelCase__ ) for block in self.transformer_blocks: UpperCAmelCase__ = block(lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) UpperCAmelCase__ = self.norm_out(lowerCamelCase__ ) if self.prd_embedding is not None: UpperCAmelCase__ = hidden_states[:, -1] else: UpperCAmelCase__ = hidden_states[:, additional_embeddings_len:] UpperCAmelCase__ = self.proj_to_clip_embeddings(lowerCamelCase__ ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
706
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ : int = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Union[str, Any] = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
632
0
"""simple docstring""" import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": lowerCAmelCase__ : Tuple = '%20'.join(argv[1:]) if len(argv) > 1 else quote(str(input('Search: '))) print('Googling.....') lowerCAmelCase__ : Tuple = F"""https://www.google.com/search?q={query}&num=100""" lowerCAmelCase__ : Union[str, Any] = requests.get( url, headers={'User-Agent': str(UserAgent().random)}, ) try: lowerCAmelCase__ : Dict = ( BeautifulSoup(res.text, 'html.parser') .find('div', attrs={'class': 'yuRUbf'}) .find('a') .get('href') ) except AttributeError: lowerCAmelCase__ : List[Any] = parse_qs( BeautifulSoup(res.text, 'html.parser') .find('div', attrs={'class': 'kCrYT'}) .find('a') .get('href') )['url'][0] webbrowser.open(link)
707
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase__ : Optional[int] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( image=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
632
0
"""simple docstring""" def a_ ( lowerCamelCase = 1_0_0_0_0_0_0 ): UpperCAmelCase__ = limit + 1 UpperCAmelCase__ = [0] * limit for first_term in range(1 , lowerCamelCase ): for n in range(lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a UpperCAmelCase__ = sum(1 for x in frequency[1:limit] if x == 1_0 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
708
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def a_ ( lowerCamelCase = 2_0_0_0_0_0_0 ): UpperCAmelCase__ = [0] UpperCAmelCase__ = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target UpperCAmelCase__ = 0 # the area corresponding to the grid that gives the product closest to target UpperCAmelCase__ = 0 # an estimate of b, using the quadratic formula UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the triangle number corresponding to b_floor UpperCAmelCase__ = 42 # the triangle number corresponding to b_ceil UpperCAmelCase__ = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): UpperCAmelCase__ = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 UpperCAmelCase__ = floor(lowerCamelCase ) UpperCAmelCase__ = ceil(lowerCamelCase ) UpperCAmelCase__ = triangle_numbers[b_floor] UpperCAmelCase__ = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_first_guess * triangle_a UpperCAmelCase__ = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_second_guess * triangle_a UpperCAmelCase__ = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
632
0
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : List[str] = logging.get_logger(__name__) lowerCAmelCase__ : List[Any] = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } lowerCAmelCase__ : Optional[int] = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } lowerCAmelCase__ : Any = {'facebook/blenderbot_small-90M': 512} def a_ ( lowerCamelCase ): UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char UpperCAmelCase__ = set(lowerCamelCase ) return pairs class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = ["input_ids", "attention_mask"] def __init__( self : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Tuple="__start__" ,lowerCamelCase__ : Tuple="__end__" ,lowerCamelCase__ : Optional[int]="__unk__" ,lowerCamelCase__ : List[Any]="__null__" ,**lowerCamelCase__ : Optional[int] ,): super().__init__(unk_token=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,**lowerCamelCase__ ) with open(lowerCamelCase__ ,encoding='utf-8' ) as vocab_handle: UpperCAmelCase__ = json.load(lowerCamelCase__ ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} with open(lowerCamelCase__ ,encoding='utf-8' ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split('\n' )[1:-1] UpperCAmelCase__ = [tuple(merge.split() ) for merge in merges] UpperCAmelCase__ = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) UpperCAmelCase__ = {} @property def __lowerCAmelCase ( self : Optional[Any] ): return len(self.encoder ) def __lowerCAmelCase ( self : str ): return dict(self.encoder ,**self.added_tokens_encoder ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : str ): if token in self.cache: return self.cache[token] UpperCAmelCase__ = re.sub('([.,!?()])' ,R' \1' ,lowerCamelCase__ ) UpperCAmelCase__ = re.sub('(\')' ,R' \1 ' ,lowerCamelCase__ ) UpperCAmelCase__ = re.sub(R'\s{2,}' ,' ' ,lowerCamelCase__ ) if "\n" in token: UpperCAmelCase__ = token.replace('\n' ,' __newln__' ) UpperCAmelCase__ = token.split(' ' ) UpperCAmelCase__ = [] for token in tokens: if not len(lowerCamelCase__ ): continue UpperCAmelCase__ = token.lower() UpperCAmelCase__ = tuple(lowerCamelCase__ ) UpperCAmelCase__ = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) UpperCAmelCase__ = get_pairs(lowerCamelCase__ ) if not pairs: words.append(lowerCamelCase__ ) continue while True: UpperCAmelCase__ = min(lowerCamelCase__ ,key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ ,float('inf' ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(lowerCamelCase__ ): try: UpperCAmelCase__ = word.index(lowerCamelCase__ ,lowerCamelCase__ ) new_word.extend(word[i:j] ) UpperCAmelCase__ = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(lowerCamelCase__ ) UpperCAmelCase__ = new_word if len(lowerCamelCase__ ) == 1: break else: UpperCAmelCase__ = get_pairs(lowerCamelCase__ ) UpperCAmelCase__ = '@@ '.join(lowerCamelCase__ ) UpperCAmelCase__ = word[:-4] UpperCAmelCase__ = word words.append(lowerCamelCase__ ) return " ".join(lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : str ): UpperCAmelCase__ = [] UpperCAmelCase__ = re.findall(R'\S+\n?' ,lowerCamelCase__ ) for token in words: split_tokens.extend(list(self.bpe(lowerCamelCase__ ).split(' ' ) ) ) return split_tokens def __lowerCAmelCase ( self : str ,lowerCamelCase__ : str ): UpperCAmelCase__ = token.lower() return self.encoder.get(lowerCamelCase__ ,self.encoder.get(self.unk_token ) ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : int ): return self.decoder.get(lowerCamelCase__ ,self.unk_token ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : List[str] ): UpperCAmelCase__ = ' '.join(lowerCamelCase__ ).replace('@@ ' ,'' ).strip() return out_string def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase__ ,ensure_ascii=lowerCamelCase__ ) + '\n' ) UpperCAmelCase__ = 0 with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' ' Please check that the tokenizer is not corrupted!' ) UpperCAmelCase__ = token_index writer.write(' '.join(lowerCamelCase__ ) + '\n' ) index += 1 return vocab_file, merge_file
709
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : Optional[Any] = ['model.decoder.embed_positions.weights'] def a_ ( lowerCamelCase ): if "emb" in name: UpperCAmelCase__ = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: UpperCAmelCase__ = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: UpperCAmelCase__ = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: UpperCAmelCase__ = name.replace('linear1' , 'fc1' ) if "linear2" in name: UpperCAmelCase__ = name.replace('linear2' , 'fc2' ) if "norm1" in name: UpperCAmelCase__ = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: UpperCAmelCase__ = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: UpperCAmelCase__ = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: UpperCAmelCase__ = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: UpperCAmelCase__ = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase__ = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = list(state_dict.keys() ) UpperCAmelCase__ = {} for key in keys: UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) UpperCAmelCase__ = rename_keys(lowerCamelCase ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase__ = val[:hidden_size, :] UpperCAmelCase__ = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase__ = val else: UpperCAmelCase__ = val return state_dict, enc_dec_proj_state_dict def a_ ( lowerCamelCase ): if checkpoint == "small": # default config values UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif checkpoint == "medium": UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 2_4 elif checkpoint == "large": UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 3_2 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) UpperCAmelCase__ = MusicgenDecoderConfig( hidden_size=lowerCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=lowerCamelCase , num_attention_heads=lowerCamelCase , ) return config @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="cpu" ): UpperCAmelCase__ = MusicGen.get_pretrained(lowerCamelCase , device=lowerCamelCase ) UpperCAmelCase__ = decoder_config_from_checkpoint(lowerCamelCase ) UpperCAmelCase__ = fairseq_model.lm.state_dict() UpperCAmelCase__ , UpperCAmelCase__ = rename_state_dict( lowerCamelCase , hidden_size=decoder_config.hidden_size ) UpperCAmelCase__ = TaEncoderModel.from_pretrained('t5-base' ) UpperCAmelCase__ = EncodecModel.from_pretrained('facebook/encodec_32khz' ) UpperCAmelCase__ = MusicgenForCausalLM(lowerCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase__ , UpperCAmelCase__ = decoder.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model UpperCAmelCase__ = MusicgenForConditionalGeneration(text_encoder=lowerCamelCase , audio_encoder=lowerCamelCase , decoder=lowerCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(lowerCamelCase ) # check we can do a forward pass UpperCAmelCase__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase__ = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits if logits.shape != (8, 1, 2_0_4_8): raise ValueError('Incorrect shape for logits' ) # now construct the processor UpperCAmelCase__ = AutoTokenizer.from_pretrained('t5-base' ) UpperCAmelCase__ = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) UpperCAmelCase__ = MusicgenProcessor(feature_extractor=lowerCamelCase , tokenizer=lowerCamelCase ) # set the appropriate bos/pad token ids UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 2_0_4_8 # set other default generation config params UpperCAmelCase__ = int(3_0 * audio_encoder.config.frame_rate ) UpperCAmelCase__ = True UpperCAmelCase__ = 3.0 if pytorch_dump_folder is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(lowerCamelCase ) processor.save_pretrained(lowerCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(lowerCamelCase ) processor.push_to_hub(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
632
0
"""simple docstring""" def a_ ( lowerCamelCase ): if length <= 0 or not isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError('Length must be a positive integer.' ) return [n * (2 * n - 1) for n in range(lowerCamelCase )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
710
"""simple docstring""" lowerCAmelCase__ : Tuple = range(2, 20 + 1) lowerCAmelCase__ : Optional[Any] = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase__ : dict[int, dict[int, list[list[int]]]] = {} def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(a_i[j] for j in range(lowerCamelCase , len(lowerCamelCase ) ) ) UpperCAmelCase__ = sum(a_i[j] * base[j] for j in range(min(len(lowerCamelCase ) , lowerCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ = 0, 0 UpperCAmelCase__ = n - i UpperCAmelCase__ = memo.get(lowerCamelCase ) if sub_memo is not None: UpperCAmelCase__ = sub_memo.get(lowerCamelCase ) if jumps is not None and len(lowerCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ = -1 for _k in range(len(lowerCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ = diff + c for j in range(min(lowerCamelCase , len(lowerCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) if new_c > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: UpperCAmelCase__ = [] else: UpperCAmelCase__ = {c: []} UpperCAmelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , k - 1 , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ = compute(lowerCamelCase , lowerCamelCase , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ = 0 while j < len(lowerCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowerCamelCase , (diff, dn, k) ) return (diff, dn) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if i >= n: return 0, i if k > len(lowerCamelCase ): a_i.extend([0 for _ in range(k - len(lowerCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 0, 0, 0 for j in range(len(lowerCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ = ds_c + ds_b diff += addend UpperCAmelCase__ = 0 for j in range(lowerCamelCase ): UpperCAmelCase__ = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return diff, i - start_i def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): for j in range(lowerCamelCase , len(lowerCamelCase ) ): UpperCAmelCase__ = digits[j] + addend if s >= 1_0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) UpperCAmelCase__ = addend // 1_0 + quotient else: UpperCAmelCase__ = s UpperCAmelCase__ = addend // 1_0 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) digits.append(lowerCamelCase ) def a_ ( lowerCamelCase = 1_0**1_5 ): UpperCAmelCase__ = [1] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , 2_0 , i + dn , lowerCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ = 0 for j in range(len(lowerCamelCase ) ): a_n += digits[j] * 1_0**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def a_ ( lowerCamelCase ): if "model" in orig_key: UpperCAmelCase__ = orig_key.replace('model.' , '' ) if "norm1" in orig_key: UpperCAmelCase__ = orig_key.replace('norm1' , 'attention.output.LayerNorm' ) if "norm2" in orig_key: UpperCAmelCase__ = orig_key.replace('norm2' , 'output.LayerNorm' ) if "norm" in orig_key: UpperCAmelCase__ = orig_key.replace('norm' , 'LayerNorm' ) if "transformer" in orig_key: UpperCAmelCase__ = orig_key.split('.' )[0].split('_' )[-1] UpperCAmelCase__ = orig_key.replace(f'''transformer_{layer_num}''' , f'''encoder.layer.{layer_num}''' ) if "mha.attn" in orig_key: UpperCAmelCase__ = orig_key.replace('mha.attn' , 'attention.self' ) if "mha" in orig_key: UpperCAmelCase__ = orig_key.replace('mha' , 'attention' ) if "W_q" in orig_key: UpperCAmelCase__ = orig_key.replace('W_q' , 'self.query' ) if "W_k" in orig_key: UpperCAmelCase__ = orig_key.replace('W_k' , 'self.key' ) if "W_v" in orig_key: UpperCAmelCase__ = orig_key.replace('W_v' , 'self.value' ) if "ff1" in orig_key: UpperCAmelCase__ = orig_key.replace('ff1' , 'intermediate.dense' ) if "ff2" in orig_key: UpperCAmelCase__ = orig_key.replace('ff2' , 'output.dense' ) if "ff" in orig_key: UpperCAmelCase__ = orig_key.replace('ff' , 'output.dense' ) if "mlm_class" in orig_key: UpperCAmelCase__ = orig_key.replace('mlm.mlm_class' , 'cls.predictions.decoder' ) if "mlm" in orig_key: UpperCAmelCase__ = orig_key.replace('mlm' , 'cls.predictions.transform' ) if "cls" not in orig_key: UpperCAmelCase__ = 'yoso.' + orig_key return orig_key def a_ ( lowerCamelCase , lowerCamelCase ): for key in orig_state_dict.copy().keys(): UpperCAmelCase__ = orig_state_dict.pop(lowerCamelCase ) if ("pooler" in key) or ("sen_class" in key): continue else: UpperCAmelCase__ = val UpperCAmelCase__ = orig_state_dict['cls.predictions.decoder.bias'] UpperCAmelCase__ = torch.arange(lowerCamelCase ).expand((1, -1) ) + 2 return orig_state_dict def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = torch.load(lowerCamelCase , map_location='cpu' )['model_state_dict'] UpperCAmelCase__ = YosoConfig.from_json_file(lowerCamelCase ) UpperCAmelCase__ = YosoForMaskedLM(lowerCamelCase ) UpperCAmelCase__ = convert_checkpoint_helper(config.max_position_embeddings , lowerCamelCase ) print(model.load_state_dict(lowerCamelCase ) ) model.eval() model.save_pretrained(lowerCamelCase ) print(f'''Checkpoint successfuly converted. Model saved at {pytorch_dump_path}''' ) if __name__ == "__main__": lowerCAmelCase__ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--pytorch_model_path', default=None, type=str, required=True, help='Path to YOSO pytorch checkpoint.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The json file for YOSO model config.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
711
"""simple docstring""" import random class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : str ): UpperCAmelCase__ = [ord(lowerCamelCase__ ) for i in text] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i in plain: UpperCAmelCase__ = random.randint(1 ,300 ) UpperCAmelCase__ = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): UpperCAmelCase__ = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ : Dict = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
632
0
"""simple docstring""" import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize('dataset_size' , [None, 4_0_0 * 2**2_0, 6_0_0 * 2**2_0] ) @pytest.mark.parametrize('input_in_memory_max_size' , ['default', 0, 1_0_0 * 2**2_0, 9_0_0 * 2**2_0] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , 'IN_MEMORY_MAX_SIZE' , lowerCamelCase ) UpperCAmelCase__ = datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: UpperCAmelCase__ = dataset_size < in_memory_max_size else: UpperCAmelCase__ = False UpperCAmelCase__ = is_small_dataset(lowerCamelCase ) assert result == expected
712
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
632
0
"""simple docstring""" def a_ ( ): return 1 def a_ ( lowerCamelCase ): return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def a_ ( lowerCamelCase ): return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowerCamelCase ) def a_ ( lowerCamelCase ): return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowerCamelCase ) def a_ ( lowerCamelCase ): return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowerCamelCase ) def a_ ( lowerCamelCase ): return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowerCamelCase ) def a_ ( lowerCamelCase ): return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowerCamelCase ) def a_ ( lowerCamelCase ): return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowerCamelCase ) def a_ ( lowerCamelCase = 2_0_0 ): return two_pound(lowerCamelCase ) if __name__ == "__main__": print(solution(int(input().strip())))
713
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Dict ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_torch def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,torch.tensor(lowerCamelCase__ ) ,torch.tensor(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) @require_vision @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_tf def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,tf.convert_to_tensor(lowerCamelCase__ ) ,tf.convert_to_tensor(lowerCamelCase__ ) ,return_tensors='tf' ,) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowerCamelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = np.random.randint(0 ,2 ,size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase__ = [tf.convert_to_tensor(lowerCamelCase__ )] UpperCAmelCase__ = [torch.tensor(lowerCamelCase__ )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='pt' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) )
632
0
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( "The RoBERTa Model transformer with early exiting (DeeRoBERTa). " , __UpperCAmelCase , ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = RobertaConfig snake_case__ = "roberta" def __init__( self : List[str] ,lowerCamelCase__ : Dict ): super().__init__(lowerCamelCase__ ) UpperCAmelCase__ = RobertaEmbeddings(lowerCamelCase__ ) self.init_weights() @add_start_docstrings( "RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top,\n also takes care of multi-layer training. " , __UpperCAmelCase , ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = RobertaConfig snake_case__ = "roberta" def __init__( self : Union[str, Any] ,lowerCamelCase__ : Optional[int] ): super().__init__(lowerCamelCase__ ) UpperCAmelCase__ = config.num_labels UpperCAmelCase__ = config.num_hidden_layers UpperCAmelCase__ = DeeRobertaModel(lowerCamelCase__ ) UpperCAmelCase__ = nn.Dropout(config.hidden_dropout_prob ) UpperCAmelCase__ = nn.Linear(config.hidden_size ,self.config.num_labels ) @add_start_docstrings_to_model_forward(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : str=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int=-1 ,lowerCamelCase__ : Any=False ,): UpperCAmelCase__ = self.num_layers try: UpperCAmelCase__ = self.roberta( lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,position_ids=lowerCamelCase__ ,head_mask=lowerCamelCase__ ,inputs_embeds=lowerCamelCase__ ,) UpperCAmelCase__ = outputs[1] UpperCAmelCase__ = self.dropout(lowerCamelCase__ ) UpperCAmelCase__ = self.classifier(lowerCamelCase__ ) UpperCAmelCase__ = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: UpperCAmelCase__ = e.message UpperCAmelCase__ = e.exit_layer UpperCAmelCase__ = outputs[0] if not self.training: UpperCAmelCase__ = entropy(lowerCamelCase__ ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] if labels is not None: if self.num_labels == 1: # We are doing regression UpperCAmelCase__ = MSELoss() UpperCAmelCase__ = loss_fct(logits.view(-1 ) ,labels.view(-1 ) ) else: UpperCAmelCase__ = CrossEntropyLoss() UpperCAmelCase__ = loss_fct(logits.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) # work with highway exits UpperCAmelCase__ = [] for highway_exit in outputs[-1]: UpperCAmelCase__ = highway_exit[0] if not self.training: highway_logits_all.append(lowerCamelCase__ ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression UpperCAmelCase__ = MSELoss() UpperCAmelCase__ = loss_fct(highway_logits.view(-1 ) ,labels.view(-1 ) ) else: UpperCAmelCase__ = CrossEntropyLoss() UpperCAmelCase__ = loss_fct(highway_logits.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) highway_losses.append(lowerCamelCase__ ) if train_highway: UpperCAmelCase__ = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: UpperCAmelCase__ = (loss,) + outputs if not self.training: UpperCAmelCase__ = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: UpperCAmelCase__ = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
714
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) lowerCAmelCase__ : str = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "ctrl" snake_case__ = ["past_key_values"] snake_case__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any ,lowerCamelCase__ : str=246_534 ,lowerCamelCase__ : List[str]=256 ,lowerCamelCase__ : Optional[int]=1_280 ,lowerCamelCase__ : Any=8_192 ,lowerCamelCase__ : int=48 ,lowerCamelCase__ : Optional[Any]=16 ,lowerCamelCase__ : Union[str, Any]=0.1 ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : List[str]=1e-6 ,lowerCamelCase__ : List[str]=0.0_2 ,lowerCamelCase__ : Tuple=True ,**lowerCamelCase__ : Optional[Any] ,): UpperCAmelCase__ = vocab_size UpperCAmelCase__ = n_positions UpperCAmelCase__ = n_embd UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = dff UpperCAmelCase__ = resid_pdrop UpperCAmelCase__ = embd_pdrop UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache super().__init__(**lowerCamelCase__ )
632
0
"""simple docstring""" import cva import numpy as np class snake_case : """simple docstring""" def __init__( self : Any ,lowerCamelCase__ : float ,lowerCamelCase__ : int ): if k in (0.0_4, 0.0_6): UpperCAmelCase__ = k UpperCAmelCase__ = window_size else: raise ValueError('invalid k value' ) def __str__( self : List[Any] ): return str(self.k ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : str ): UpperCAmelCase__ = cva.imread(lowerCamelCase__ ,0 ) UpperCAmelCase__ , UpperCAmelCase__ = img.shape UpperCAmelCase__ = [] UpperCAmelCase__ = img.copy() UpperCAmelCase__ = cva.cvtColor(lowerCamelCase__ ,cva.COLOR_GRAY2RGB ) UpperCAmelCase__ , UpperCAmelCase__ = np.gradient(lowerCamelCase__ ) UpperCAmelCase__ = dx**2 UpperCAmelCase__ = dy**2 UpperCAmelCase__ = dx * dy UpperCAmelCase__ = 0.0_4 UpperCAmelCase__ = self.window_size // 2 for y in range(lowerCamelCase__ ,h - offset ): for x in range(lowerCamelCase__ ,w - offset ): UpperCAmelCase__ = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() UpperCAmelCase__ = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() UpperCAmelCase__ = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() UpperCAmelCase__ = (wxx * wyy) - (wxy**2) UpperCAmelCase__ = wxx + wyy UpperCAmelCase__ = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) ,0 ) color_img.itemset((y, x, 1) ,0 ) color_img.itemset((y, x, 2) ,255 ) return color_img, corner_list if __name__ == "__main__": lowerCAmelCase__ : List[str] = HarrisCorner(0.04, 3) lowerCAmelCase__ : Optional[Any] = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
715
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCAmelCase__ : Dict = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def a_ ( ): UpperCAmelCase__ = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCAmelCase__ = get_sagemaker_input() else: UpperCAmelCase__ = get_cluster_input() return config def a_ ( lowerCamelCase=None ): if subparsers is not None: UpperCAmelCase__ = subparsers.add_parser('config' , description=lowerCamelCase ) else: UpperCAmelCase__ = argparse.ArgumentParser('Accelerate config command' , description=lowerCamelCase ) parser.add_argument( '--config_file' , default=lowerCamelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def a_ ( lowerCamelCase ): UpperCAmelCase__ = get_user_input() if args.config_file is not None: UpperCAmelCase__ = args.config_file else: if not os.path.isdir(lowerCamelCase ): os.makedirs(lowerCamelCase ) UpperCAmelCase__ = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(lowerCamelCase ) else: config.to_yaml_file(lowerCamelCase ) print(f'''accelerate configuration saved at {config_file}''' ) def a_ ( ): UpperCAmelCase__ = config_command_parser() UpperCAmelCase__ = parser.parse_args() config_command(lowerCamelCase ) if __name__ == "__main__": main()
632
0
"""simple docstring""" from __future__ import annotations lowerCAmelCase__ : Optional[Any] = 1.6021E-19 # units = C def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , ): if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('You cannot supply more or less than 2 values' ) elif conductivity < 0: raise ValueError('Conductivity cannot be negative' ) elif electron_conc < 0: raise ValueError('Electron concentration cannot be negative' ) elif mobility < 0: raise ValueError('mobility cannot be negative' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
716
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): return x if y == 0 else greatest_common_divisor(lowerCamelCase , x % y ) def a_ ( lowerCamelCase , lowerCamelCase ): return (x * y) // greatest_common_divisor(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase = 2_0 ): UpperCAmelCase__ = 1 for i in range(1 , n + 1 ): UpperCAmelCase__ = lcm(lowerCamelCase , lowerCamelCase ) return g if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" from math import pi def a_ ( lowerCamelCase , lowerCamelCase ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
717
"""simple docstring""" import warnings from functools import wraps from typing import Callable def a_ ( lowerCamelCase ): @wraps(lowerCamelCase ) def _inner_fn(*lowerCamelCase , **lowerCamelCase ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , lowerCamelCase , ) return fn(*lowerCamelCase , **lowerCamelCase ) return _inner_fn
632
0
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
718
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCAmelCase__ : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCAmelCase__ : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowerCAmelCase__ : set[int] = {ord(char) for char in VALID_CHARS} lowerCAmelCase__ : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = "" UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 for keychar, cipherchar in zip(cycle(lowerCamelCase ) , lowerCamelCase ): UpperCAmelCase__ = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCamelCase ) return decoded def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] for key in product(lowerCamelCase , repeat=3 ): UpperCAmelCase__ = try_key(lowerCamelCase , lowerCamelCase ) if encoded is not None: possibles.append(lowerCamelCase ) return possibles def a_ ( lowerCamelCase , lowerCamelCase ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCamelCase = "p059_cipher.txt" ): UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = Path(lowerCamelCase ).parent.joinpath(lowerCamelCase ).read_text(encoding='utf-8' ) UpperCAmelCase__ = [int(lowerCamelCase ) for number in data.strip().split(',' )] UpperCAmelCase__ = filter_valid_chars(lowerCamelCase ) for common_word in COMMON_WORDS: UpperCAmelCase__ = filter_common_word(lowerCamelCase , lowerCamelCase ) if len(lowerCamelCase ) == 1: break UpperCAmelCase__ = possibles[0] return sum(ord(lowerCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
632
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase__ : str = { 'naver-clova-ix/donut-base': 'https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json', # See all Donut models at https://huggingface.co/models?filter=donut-swin } class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "donut-swin" snake_case__ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : Tuple ,lowerCamelCase__ : int=224 ,lowerCamelCase__ : Union[str, Any]=4 ,lowerCamelCase__ : List[Any]=3 ,lowerCamelCase__ : str=96 ,lowerCamelCase__ : List[str]=[2, 2, 6, 2] ,lowerCamelCase__ : List[str]=[3, 6, 12, 24] ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Any=4.0 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=0.0 ,lowerCamelCase__ : List[Any]=0.0 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : int="gelu" ,lowerCamelCase__ : Optional[Any]=False ,lowerCamelCase__ : Any=0.0_2 ,lowerCamelCase__ : List[str]=1e-5 ,**lowerCamelCase__ : str ,): super().__init__(**lowerCamelCase__ ) UpperCAmelCase__ = image_size UpperCAmelCase__ = patch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = embed_dim UpperCAmelCase__ = depths UpperCAmelCase__ = len(lowerCamelCase__ ) UpperCAmelCase__ = num_heads UpperCAmelCase__ = window_size UpperCAmelCase__ = mlp_ratio UpperCAmelCase__ = qkv_bias UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = drop_path_rate UpperCAmelCase__ = hidden_act UpperCAmelCase__ = use_absolute_embeddings UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase__ = int(embed_dim * 2 ** (len(lowerCamelCase__ ) - 1) )
719
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = XLMProphetNetTokenizer snake_case__ = False snake_case__ = True def __lowerCAmelCase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = '[PAD]' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'[PAD]' ) self.assertEqual(vocab_keys[1] ,'[CLS]' ) self.assertEqual(vocab_keys[-1] ,'j' ) self.assertEqual(len(lowerCamelCase__ ) ,1_012 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_012 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) UpperCAmelCase__ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] ,) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] ,) @cached_property def __lowerCAmelCase ( self : Dict ): return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'Hello World!' UpperCAmelCase__ = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/xprophetnet-large-wiki100-cased' ,revision='1acad1643ddd54a44df6a1b797ada8373685d90e' ,)
632
0
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = IFPipeline snake_case__ = TEXT_TO_IMAGE_PARAMS - {"width", "height", "latents"} snake_case__ = TEXT_TO_IMAGE_BATCH_PARAMS snake_case__ = PipelineTesterMixin.required_optional_params - {"latents"} def __lowerCAmelCase ( self : Any ): return self._get_dummy_components() def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Dict=0 ): if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : List[Any] ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' ,reason='float16 requires CUDA' ) def __lowerCAmelCase ( self : List[Any] ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def __lowerCAmelCase ( self : Tuple ): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def __lowerCAmelCase ( self : Tuple ): self._test_save_load_local() def __lowerCAmelCase ( self : Dict ): self._test_inference_batch_single_identical( expected_max_diff=1e-2 ,) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() ,reason='XFormers attention is only available with CUDA and `xformers` installed' ,) def __lowerCAmelCase ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Union[str, Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : List[str] ): # if UpperCAmelCase__ = IFPipeline.from_pretrained('DeepFloyd/IF-I-XL-v1.0' ,variant='fp16' ,torch_dtype=torch.floataa ) UpperCAmelCase__ = IFSuperResolutionPipeline.from_pretrained( 'DeepFloyd/IF-II-L-v1.0' ,variant='fp16' ,torch_dtype=torch.floataa ,text_encoder=lowerCamelCase__ ,tokenizer=lowerCamelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to('cuda' ) UpperCAmelCase__ , UpperCAmelCase__ = pipe_a.encode_prompt('anime turtle' ,device='cuda' ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() UpperCAmelCase__ = None UpperCAmelCase__ = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img UpperCAmelCase__ = IFImgaImgPipeline(**pipe_a.components ) UpperCAmelCase__ = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting UpperCAmelCase__ = IFInpaintingPipeline(**pipe_a.components ) UpperCAmelCase__ = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Dict ,lowerCamelCase__ : List[str] ): # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,num_inference_steps=2 ,generator=lowerCamelCase__ ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (64, 64, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,image=lowerCamelCase__ ,generator=lowerCamelCase__ ,num_inference_steps=2 ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (256, 256, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Dict ): # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,image=lowerCamelCase__ ,num_inference_steps=2 ,generator=lowerCamelCase__ ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (64, 64, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = floats_tensor((1, 3, 256, 256) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,image=lowerCamelCase__ ,original_image=lowerCamelCase__ ,generator=lowerCamelCase__ ,num_inference_steps=2 ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (256, 256, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(1 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,image=lowerCamelCase__ ,mask_image=lowerCamelCase__ ,num_inference_steps=2 ,generator=lowerCamelCase__ ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (64, 64, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = floats_tensor((1, 3, 64, 64) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = floats_tensor((1, 3, 256, 256) ,rng=random.Random(0 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = floats_tensor((1, 3, 256, 256) ,rng=random.Random(1 ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = pipe_a( prompt_embeds=lowerCamelCase__ ,negative_prompt_embeds=lowerCamelCase__ ,image=lowerCamelCase__ ,mask_image=lowerCamelCase__ ,original_image=lowerCamelCase__ ,generator=lowerCamelCase__ ,num_inference_steps=2 ,output_type='np' ,) UpperCAmelCase__ = output.images[0] assert image.shape == (256, 256, 3) UpperCAmelCase__ = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 UpperCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy' ) assert_mean_pixel_difference(lowerCamelCase__ ,lowerCamelCase__ ) def a_ ( ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
720
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def a_ ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def a_ ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] lowerCAmelCase__ : str = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') lowerCAmelCase__ : str = sorted({word.strip().lower() for word in data.splitlines()}) lowerCAmelCase__ : Optional[Any] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": lowerCAmelCase__ : int = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
632
0
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Any=13 ,lowerCamelCase__ : List[Any]=7 ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Tuple=99 ,lowerCamelCase__ : str=32 ,lowerCamelCase__ : Optional[Any]=5 ,lowerCamelCase__ : Union[str, Any]=4 ,lowerCamelCase__ : Dict=37 ,lowerCamelCase__ : Optional[int]="gelu" ,lowerCamelCase__ : Optional[int]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : List[Any]=512 ,lowerCamelCase__ : List[Any]=16 ,lowerCamelCase__ : Union[str, Any]=2 ,lowerCamelCase__ : Tuple=0.0_2 ,lowerCamelCase__ : int=4 ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_choices def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ = None if self.use_attention_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase__ = RoFormerConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowerCamelCase__ ,initializer_range=self.initializer_range ,) return config, input_ids, token_type_ids, attention_mask def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = True snake_case__ = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = FlaxRoFormerModelTester(self ) @slow def __lowerCAmelCase ( self : Any ): for model_class_name in self.all_model_classes: UpperCAmelCase__ = model_class_name.from_pretrained('junnyu/roformer_chinese_small' ,from_pt=lowerCamelCase__ ) UpperCAmelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class snake_case ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = FlaxRoFormerForMaskedLM.from_pretrained('junnyu/roformer_chinese_base' ) UpperCAmelCase__ = jnp.array([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ = model(lowerCamelCase__ )[0] UpperCAmelCase__ = 50_000 UpperCAmelCase__ = (1, 6, vocab_size) self.assertEqual(output.shape ,lowerCamelCase__ ) UpperCAmelCase__ = jnp.array( [[[-0.1_2_0_5, -1.0_2_6_5, 0.2_9_2_2], [-1.5_1_3_4, 0.1_9_7_4, 0.1_5_1_9], [-5.0_1_3_5, -3.9_0_0_3, -0.8_4_0_4]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] ,lowerCamelCase__ ,atol=1e-4 ) )
721
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class snake_case ( ctypes.Structure ): """simple docstring""" snake_case__ = [("size", ctypes.c_int), ("visible", ctypes.c_byte)] def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def a_ ( ): try: hide_cursor() yield finally: show_cursor()
632
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) lowerCAmelCase__ : str = '▁' lowerCAmelCase__ : str = {'vocab_file': 'sentencepiece.bpe.model'} lowerCAmelCase__ : Any = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } lowerCAmelCase__ : List[str] = { 'facebook/mbart-large-en-ro': 1_024, 'facebook/mbart-large-cc25': 1_024, } # fmt: off lowerCAmelCase__ : Optional[int] = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = ["input_ids", "attention_mask"] snake_case__ = [] snake_case__ = [] def __init__( self : Any ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[Any]="<s>" ,lowerCamelCase__ : str="</s>" ,lowerCamelCase__ : Union[str, Any]="</s>" ,lowerCamelCase__ : Optional[int]="<s>" ,lowerCamelCase__ : str="<unk>" ,lowerCamelCase__ : Optional[int]="<pad>" ,lowerCamelCase__ : Dict="<mask>" ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Any=None ,lowerCamelCase__ : Optional[Dict[str, Any]] = None ,lowerCamelCase__ : Optional[int]=None ,**lowerCamelCase__ : Any ,): # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,src_lang=lowerCamelCase__ ,tgt_lang=lowerCamelCase__ ,additional_special_tokens=lowerCamelCase__ ,sp_model_kwargs=self.sp_model_kwargs ,**lowerCamelCase__ ,) UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) UpperCAmelCase__ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase__ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase__ = 1 UpperCAmelCase__ = len(self.sp_model ) UpperCAmelCase__ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCamelCase__ ) } UpperCAmelCase__ = {v: k for k, v in self.lang_code_to_id.items()} UpperCAmelCase__ = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCAmelCase__ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCAmelCase__ = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCAmelCase__ = src_lang if src_lang is not None else 'en_XX' UpperCAmelCase__ = self.lang_code_to_id[self._src_lang] UpperCAmelCase__ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : Optional[Any] ): UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None UpperCAmelCase__ = self.sp_model.serialized_model_proto() return state def __setstate__( self : Any ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def __lowerCAmelCase ( self : str ): return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __lowerCAmelCase ( self : List[Any] ): return self._src_lang @src_lang.setter def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : str ): UpperCAmelCase__ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ,lowerCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ ,token_ids_a=lowerCamelCase__ ,already_has_special_tokens=lowerCamelCase__ ) UpperCAmelCase__ = [1] * len(self.prefix_tokens ) UpperCAmelCase__ = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCamelCase__ )) + suffix_ones return prefix_ones + ([0] * len(lowerCamelCase__ )) + ([0] * len(lowerCamelCase__ )) + suffix_ones def __lowerCAmelCase ( self : str ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowerCAmelCase ( self : int ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] ,lowerCamelCase__ : Optional[str] ,**lowerCamelCase__ : Any ): if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) UpperCAmelCase__ = src_lang UpperCAmelCase__ = self(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) UpperCAmelCase__ = self.convert_tokens_to_ids(lowerCamelCase__ ) UpperCAmelCase__ = tgt_lang_id return inputs def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : str ): return self.sp_model.encode(lowerCamelCase__ ,out_type=lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : Optional[Any] ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase__ = self.sp_model.PieceToId(lowerCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : List[Any] ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = ''.join(lowerCamelCase__ ).replace(lowerCamelCase__ ,' ' ).strip() return out_string def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ ,'wb' ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : str = "en_XX" ,lowerCamelCase__ : Optional[List[str]] = None ,lowerCamelCase__ : str = "ro_RO" ,**lowerCamelCase__ : Union[str, Any] ,): UpperCAmelCase__ = src_lang UpperCAmelCase__ = tgt_lang return super().prepare_seqaseq_batch(lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): return self.set_src_lang_special_tokens(self.src_lang ) def __lowerCAmelCase ( self : List[Any] ): return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : List[Any] ): UpperCAmelCase__ = self.lang_code_to_id[src_lang] UpperCAmelCase__ = [] UpperCAmelCase__ = [self.eos_token_id, self.cur_lang_code] def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : str ): UpperCAmelCase__ = self.lang_code_to_id[lang] UpperCAmelCase__ = [] UpperCAmelCase__ = [self.eos_token_id, self.cur_lang_code]
700
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class snake_case ( __UpperCAmelCase ): """simple docstring""" def __get__( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) UpperCAmelCase__ = '__cached_' + self.fget.__name__ UpperCAmelCase__ = getattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) if cached is None: UpperCAmelCase__ = self.fget(lowerCamelCase__ ) setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return cached def a_ ( lowerCamelCase ): UpperCAmelCase__ = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def a_ ( lowerCamelCase ): if is_torch_fx_proxy(lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return _is_numpy(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.Tensor ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.device ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_device(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch if isinstance(lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) else: return False return isinstance(lowerCamelCase , torch.dtype ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_dtype(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf return isinstance(lowerCamelCase , tf.Tensor ) def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tensorflow(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowerCamelCase , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowerCamelCase ) return type(lowerCamelCase ) == tf.Tensor def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowerCamelCase ) def a_ ( lowerCamelCase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowerCamelCase , jnp.ndarray ) def a_ ( lowerCamelCase ): return False if not is_flax_available() else _is_jax(lowerCamelCase ) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return [to_py_obj(lowerCamelCase ) for o in obj] elif is_tf_tensor(lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ).tolist() elif isinstance(lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return np.array(lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): return obj.numpy() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ) else: return obj class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = fields(self ) # Safety and consistency checks if not len(lowerCamelCase__ ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) UpperCAmelCase__ = getattr(self ,class_fields[0].name ) UpperCAmelCase__ = all(getattr(self ,field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase__ ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = first_field.items() UpperCAmelCase__ = True else: try: UpperCAmelCase__ = iter(lowerCamelCase__ ) UpperCAmelCase__ = True except TypeError: UpperCAmelCase__ = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase__ ): if ( not isinstance(lowerCamelCase__ ,(list, tuple) ) or not len(lowerCamelCase__ ) == 2 or not isinstance(element[0] ,lowerCamelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute UpperCAmelCase__ = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self ,element[0] ,element[1] ) if element[1] is not None: UpperCAmelCase__ = element[1] elif first_field is not None: UpperCAmelCase__ = first_field else: for field in class_fields: UpperCAmelCase__ = getattr(self ,field.name ) if v is not None: UpperCAmelCase__ = v def __delitem__( self : List[Any] ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : Dict ): raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : int ,**lowerCamelCase__ : Optional[Any] ): raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : List[Any] ,*lowerCamelCase__ : int ,**lowerCamelCase__ : List[Any] ): raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Optional[int] ): raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[Any] ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __setitem__( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): # Will raise a KeyException if needed super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): return tuple(self[k] for k in self.keys() ) class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : str ,lowerCamelCase__ : Optional[int] ): raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "longest" snake_case__ = "max_length" snake_case__ = "do_not_pad" class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "pt" snake_case__ = "tf" snake_case__ = "np" snake_case__ = "jax" class snake_case : """simple docstring""" def __init__( self : int ,lowerCamelCase__ : List[ContextManager] ): UpperCAmelCase__ = context_managers UpperCAmelCase__ = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase__ ) def __exit__( self : List[Any] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Dict ): self.stack.__exit__(*lowerCamelCase__ ,**lowerCamelCase__ ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( lowerCamelCase ): UpperCAmelCase__ = model_class.__name__ UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( lowerCamelCase , lowerCamelCase = "" , lowerCamelCase = "." ): def _flatten_dict(lowerCamelCase , lowerCamelCase="" , lowerCamelCase="." ): for k, v in d.items(): UpperCAmelCase__ = str(lowerCamelCase ) + delimiter + str(lowerCamelCase ) if parent_key else k if v and isinstance(lowerCamelCase , lowerCamelCase ): yield from flatten_dict(lowerCamelCase , lowerCamelCase , delimiter=lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) @contextmanager def a_ ( lowerCamelCase , lowerCamelCase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.transpose(lowerCamelCase , axes=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.T if axes is None else array.permute(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.transpose(lowerCamelCase , perm=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.transpose(lowerCamelCase , axes=lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.reshape(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.reshape(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.reshape(lowerCamelCase , lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.reshape(lowerCamelCase , lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.squeeze(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.expand_dims(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.unsqueeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.expand_dims(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.expand_dims(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.size(lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.numel() elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.size(lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): for key, value in auto_map.items(): if isinstance(lowerCamelCase , (tuple, list) ): UpperCAmelCase__ = [f'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: UpperCAmelCase__ = f'''{repo_id}--{value}''' return auto_map def a_ ( lowerCamelCase ): for base_class in inspect.getmro(lowerCamelCase ): UpperCAmelCase__ = base_class.__module__ UpperCAmelCase__ = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
632
0
"""simple docstring""" import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a_ ( lowerCamelCase ): return EnvironmentCommand() class snake_case ( __UpperCAmelCase ): """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : ArgumentParser ): UpperCAmelCase__ = parser.add_parser('env' ) download_parser.set_defaults(func=lowerCamelCase__ ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = huggingface_hub.__version__ UpperCAmelCase__ = 'not installed' UpperCAmelCase__ = 'NA' if is_torch_available(): import torch UpperCAmelCase__ = torch.__version__ UpperCAmelCase__ = torch.cuda.is_available() UpperCAmelCase__ = 'not installed' if is_transformers_available(): import transformers UpperCAmelCase__ = transformers.__version__ UpperCAmelCase__ = 'not installed' if is_accelerate_available(): import accelerate UpperCAmelCase__ = accelerate.__version__ UpperCAmelCase__ = 'not installed' if is_xformers_available(): import xformers UpperCAmelCase__ = xformers.__version__ UpperCAmelCase__ = { '`diffusers` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'PyTorch version (GPU?)': f'''{pt_version} ({pt_cuda_available})''', 'Huggingface_hub version': hub_version, 'Transformers version': transformers_version, 'Accelerate version': accelerate_version, 'xFormers version': xformers_version, 'Using GPU in script?': '<fill in>', 'Using distributed or parallel set-up in script?': '<fill in>', } print('\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n' ) print(self.format_dict(lowerCamelCase__ ) ) return info @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : Any ): return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
701
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ : Union[str, Any] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained(lowerCamelCase__ ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = generator.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'cyberpunk 2077' UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = 'A painting of a squirrel eating a burger ' UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.text_to_image( prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = pipe.image_variation(lowerCamelCase__ ,generator=lowerCamelCase__ ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
632
0
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase__ ,'hidden_sizes' ) ) self.parent.assertTrue(hasattr(lowerCamelCase__ ,'num_attention_heads' ) ) class snake_case : """simple docstring""" def __init__( self : List[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Optional[int]=64 ,lowerCamelCase__ : List[Any]=3 ,lowerCamelCase__ : Union[str, Any]=3 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : Optional[Any]=1 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : Any=[128, 256, 384] ,lowerCamelCase__ : Dict=[4, 6, 8] ,lowerCamelCase__ : int=[2, 3, 4] ,lowerCamelCase__ : List[str]=[16, 16, 16] ,lowerCamelCase__ : Tuple=0 ,lowerCamelCase__ : Optional[int]=[2, 2, 2] ,lowerCamelCase__ : int=[2, 2, 2] ,lowerCamelCase__ : str=0.0_2 ,lowerCamelCase__ : str=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=2 ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = kernel_size UpperCAmelCase__ = stride UpperCAmelCase__ = padding UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = depths UpperCAmelCase__ = key_dim UpperCAmelCase__ = drop_path_rate UpperCAmelCase__ = patch_size UpperCAmelCase__ = attention_ratio UpperCAmelCase__ = mlp_ratio UpperCAmelCase__ = initializer_range UpperCAmelCase__ = [ ['Subsample', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['Subsample', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = num_labels UpperCAmelCase__ = initializer_range def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] ,self.num_labels ) UpperCAmelCase__ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self : Dict ): return LevitConfig( image_size=self.image_size ,num_channels=self.num_channels ,kernel_size=self.kernel_size ,stride=self.stride ,padding=self.padding ,patch_size=self.patch_size ,hidden_sizes=self.hidden_sizes ,num_attention_heads=self.num_attention_heads ,depths=self.depths ,key_dim=self.key_dim ,drop_path_rate=self.drop_path_rate ,mlp_ratio=self.mlp_ratio ,attention_ratio=self.attention_ratio ,initializer_range=self.initializer_range ,down_ops=self.down_ops ,) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Any ,lowerCamelCase__ : Any ): UpperCAmelCase__ = LevitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCAmelCase__ = model(lowerCamelCase__ ) UpperCAmelCase__ = (self.image_size, self.image_size) UpperCAmelCase__ , UpperCAmelCase__ = image_size[0], image_size[1] for _ in range(4 ): UpperCAmelCase__ = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) UpperCAmelCase__ = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) ,) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = LevitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCAmelCase__ = model(lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) snake_case__ = ( { "feature-extraction": LevitModel, "image-classification": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = LevitModelTester(self ) UpperCAmelCase__ = ConfigTester(self ,config_class=lowerCamelCase__ ,has_text_modality=lowerCamelCase__ ,hidden_size=37 ) def __lowerCAmelCase ( self : List[Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self : int ): return @unittest.skip(reason='Levit does not use inputs_embeds' ) def __lowerCAmelCase ( self : List[Any] ): pass @unittest.skip(reason='Levit does not support input and output embeddings' ) def __lowerCAmelCase ( self : Tuple ): pass @unittest.skip(reason='Levit does not output attentions' ) def __lowerCAmelCase ( self : Union[str, Any] ): pass def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(lowerCamelCase__ ) UpperCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ['pixel_values'] self.assertListEqual(arg_names[:1] ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): def check_hidden_states_output(lowerCamelCase__ : Tuple ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : int ): UpperCAmelCase__ = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): UpperCAmelCase__ = model(**self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) ) UpperCAmelCase__ = outputs.hidden_states UpperCAmelCase__ = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowerCamelCase__ ) ,lowerCamelCase__ ) UpperCAmelCase__ = (self.model_tester.image_size, self.model_tester.image_size) UpperCAmelCase__ , UpperCAmelCase__ = image_size[0], image_size[1] for _ in range(4 ): UpperCAmelCase__ = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) UpperCAmelCase__ = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[ height * width, self.model_tester.hidden_sizes[0], ] ,) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self : Dict ): pass def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any]=False ): UpperCAmelCase__ = super()._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): if not self.model_tester.is_training: return UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCamelCase__ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue UpperCAmelCase__ = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.train() UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) UpperCAmelCase__ = model(**lowerCamelCase__ ).loss loss.backward() def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase__ = False UpperCAmelCase__ = True for model_class in self.all_model_classes: if model_class in get_values(lowerCamelCase__ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue UpperCAmelCase__ = model_class(lowerCamelCase__ ) model.gradient_checkpointing_enable() model.to(lowerCamelCase__ ) model.train() UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) UpperCAmelCase__ = model(**lowerCamelCase__ ).loss loss.backward() def __lowerCAmelCase ( self : str ): UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = [ {'title': 'multi_label_classification', 'num_labels': 2, 'dtype': torch.float}, {'title': 'single_label_classification', 'num_labels': 1, 'dtype': torch.long}, {'title': 'regression', 'num_labels': 1, 'dtype': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCamelCase__ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f'''Testing {model_class} with {problem_type["title"]}''' ): UpperCAmelCase__ = problem_type['title'] UpperCAmelCase__ = problem_type['num_labels'] UpperCAmelCase__ = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.train() UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) if problem_type["num_labels"] > 1: UpperCAmelCase__ = inputs['labels'].unsqueeze(1 ).repeat(1 ,problem_type['num_labels'] ) UpperCAmelCase__ = inputs['labels'].to(problem_type['dtype'] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCamelCase__ ) as warning_list: UpperCAmelCase__ = model(**lowerCamelCase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def __lowerCAmelCase ( self : int ): for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = LevitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def a_ ( ): UpperCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class snake_case ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self : Dict ): return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowerCamelCase__ ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=lowerCamelCase__ ,return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): UpperCAmelCase__ = model(**lowerCamelCase__ ) # verify the logits UpperCAmelCase__ = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase__ ) UpperCAmelCase__ = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase__ ,atol=1e-4 ) )
702
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
632
0
"""simple docstring""" from datetime import datetime as dt import os from github import Github lowerCAmelCase__ : Optional[int] = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def a_ ( ): UpperCAmelCase__ = Github(os.environ['GITHUB_TOKEN'] ) UpperCAmelCase__ = g.get_repo('huggingface/transformers' ) UpperCAmelCase__ = repo.get_issues(state='open' ) for issue in open_issues: UpperCAmelCase__ = sorted([comment for comment in issue.get_comments()] , key=lambda lowerCamelCase : i.created_at , reverse=lowerCamelCase ) UpperCAmelCase__ = comments[0] if len(lowerCamelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='closed' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
703
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
632
0
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def a_ ( ): UpperCAmelCase__ = 'https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png' UpperCAmelCase__ = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ).convert('RGB' ) return image def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] # fmt: off # vision encoder rename_keys.append(('visual_encoder.cls_token', 'vision_model.embeddings.class_embedding') ) rename_keys.append(('visual_encoder.pos_embed', 'vision_model.embeddings.position_embedding') ) rename_keys.append(('visual_encoder.patch_embed.proj.weight', 'vision_model.embeddings.patch_embedding.weight') ) rename_keys.append(('visual_encoder.patch_embed.proj.bias', 'vision_model.embeddings.patch_embedding.bias') ) rename_keys.append(('ln_vision.weight', 'vision_model.post_layernorm.weight') ) rename_keys.append(('ln_vision.bias', 'vision_model.post_layernorm.bias') ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((f'''visual_encoder.blocks.{i}.norm1.weight''', f'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.norm1.bias''', f'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.norm2.weight''', f'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.norm2.bias''', f'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.attn.qkv.weight''', f'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.attn.proj.weight''', f'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((f'''visual_encoder.blocks.{i}.attn.proj.bias''', f'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.mlp.fc1.weight''', f'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.mlp.fc1.bias''', f'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.mlp.fc2.weight''', f'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((f'''visual_encoder.blocks.{i}.mlp.fc2.bias''', f'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(('Qformer.bert.embeddings.LayerNorm.weight', 'qformer.layernorm.weight') ) rename_keys.append(('Qformer.bert.embeddings.LayerNorm.bias', 'qformer.layernorm.bias') ) # fmt: on return rename_keys def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = dct.pop(lowerCamelCase ) UpperCAmelCase__ = val def a_ ( lowerCamelCase , lowerCamelCase ): for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases UpperCAmelCase__ = state_dict.pop(f'''visual_encoder.blocks.{i}.attn.q_bias''' ) UpperCAmelCase__ = state_dict.pop(f'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict UpperCAmelCase__ = torch.cat((q_bias, torch.zeros_like(lowerCamelCase , requires_grad=lowerCamelCase ), v_bias) ) UpperCAmelCase__ = qkv_bias def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = 3_6_4 if 'coco' in model_name else 2_2_4 UpperCAmelCase__ = BlipaVisionConfig(image_size=lowerCamelCase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: UpperCAmelCase__ = OPTConfig.from_pretrained('facebook/opt-2.7b' , eos_token_id=lowerCamelCase ).to_dict() elif "opt-6.7b" in model_name: UpperCAmelCase__ = OPTConfig.from_pretrained('facebook/opt-6.7b' , eos_token_id=lowerCamelCase ).to_dict() elif "t5-xl" in model_name: UpperCAmelCase__ = TaConfig.from_pretrained('google/flan-t5-xl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: UpperCAmelCase__ = TaConfig.from_pretrained('google/flan-t5-xxl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() UpperCAmelCase__ = BlipaConfig(vision_config=lowerCamelCase , text_config=lowerCamelCase ) return config, image_size @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=False ): UpperCAmelCase__ = ( AutoTokenizer.from_pretrained('facebook/opt-2.7b' ) if 'opt' in model_name else AutoTokenizer.from_pretrained('google/flan-t5-xl' ) ) UpperCAmelCase__ = tokenizer('\n' , add_special_tokens=lowerCamelCase ).input_ids[0] UpperCAmelCase__ , UpperCAmelCase__ = get_blipa_config(lowerCamelCase , eos_token_id=lowerCamelCase ) UpperCAmelCase__ = BlipaForConditionalGeneration(lowerCamelCase ).eval() UpperCAmelCase__ = { 'blip2-opt-2.7b': ('blip2_opt', 'pretrain_opt2.7b'), 'blip2-opt-6.7b': ('blip2_opt', 'pretrain_opt6.7b'), 'blip2-opt-2.7b-coco': ('blip2_opt', 'caption_coco_opt2.7b'), 'blip2-opt-6.7b-coco': ('blip2_opt', 'caption_coco_opt6.7b'), 'blip2-flan-t5-xl': ('blip2_t5', 'pretrain_flant5xl'), 'blip2-flan-t5-xl-coco': ('blip2_t5', 'caption_coco_flant5xl'), 'blip2-flan-t5-xxl': ('blip2_t5', 'pretrain_flant5xxl'), } UpperCAmelCase__ , UpperCAmelCase__ = model_name_to_original[model_name] # load original model print('Loading original model...' ) UpperCAmelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = load_model_and_preprocess( name=lowerCamelCase , model_type=lowerCamelCase , is_eval=lowerCamelCase , device=lowerCamelCase ) original_model.eval() print('Done!' ) # update state dict keys UpperCAmelCase__ = original_model.state_dict() UpperCAmelCase__ = create_rename_keys(lowerCamelCase ) for src, dest in rename_keys: rename_key(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) if key.startswith('Qformer.bert' ): UpperCAmelCase__ = key.replace('Qformer.bert' , 'qformer' ) if "attention.self" in key: UpperCAmelCase__ = key.replace('self' , 'attention' ) if "opt_proj" in key: UpperCAmelCase__ = key.replace('opt_proj' , 'language_projection' ) if "t5_proj" in key: UpperCAmelCase__ = key.replace('t5_proj' , 'language_projection' ) if key.startswith('opt' ): UpperCAmelCase__ = key.replace('opt' , 'language' ) if key.startswith('t5' ): UpperCAmelCase__ = key.replace('t5' , 'language' ) UpperCAmelCase__ = val # read in qv biases read_in_q_v_bias(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ , UpperCAmelCase__ = hf_model.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) assert len(lowerCamelCase ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] UpperCAmelCase__ = load_demo_image() UpperCAmelCase__ = vis_processors['eval'](lowerCamelCase ).unsqueeze(0 ).to(lowerCamelCase ) UpperCAmelCase__ = tokenizer(['\n'] , return_tensors='pt' ).input_ids.to(lowerCamelCase ) # create processor UpperCAmelCase__ = BlipImageProcessor( size={'height': image_size, 'width': image_size} , image_mean=lowerCamelCase , image_std=lowerCamelCase ) UpperCAmelCase__ = BlipaProcessor(image_processor=lowerCamelCase , tokenizer=lowerCamelCase ) UpperCAmelCase__ = processor(images=lowerCamelCase , return_tensors='pt' ).pixel_values.to(lowerCamelCase ) # make sure processor creates exact same pixel values assert torch.allclose(lowerCamelCase , lowerCamelCase ) original_model.to(lowerCamelCase ) hf_model.to(lowerCamelCase ) with torch.no_grad(): if "opt" in model_name: UpperCAmelCase__ = original_model({'image': original_pixel_values, 'text_input': ['']} ).logits UpperCAmelCase__ = hf_model(lowerCamelCase , lowerCamelCase ).logits else: UpperCAmelCase__ = original_model( {'image': original_pixel_values, 'text_input': ['\n'], 'text_output': ['\n']} ).logits UpperCAmelCase__ = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -1_0_0 ) UpperCAmelCase__ = hf_model(lowerCamelCase , lowerCamelCase , labels=lowerCamelCase ).logits assert original_logits.shape == logits.shape print('First values of original logits:' , original_logits[0, :3, :3] ) print('First values of HF logits:' , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": UpperCAmelCase__ = torch.tensor( [[-41.5850, -4.4440, -8.9922], [-47.4322, -5.9143, -1.7340]] , device=lowerCamelCase ) assert torch.allclose(logits[0, :3, :3] , lowerCamelCase , atol=1e-4 ) elif model_name == "blip2-flan-t5-xl-coco": UpperCAmelCase__ = torch.tensor( [[-57.0109, -9.8967, -12.6280], [-68.6578, -12.7191, -10.5065]] , device=lowerCamelCase ) else: # cast to same type UpperCAmelCase__ = logits.dtype assert torch.allclose(original_logits.to(lowerCamelCase ) , lowerCamelCase , atol=1e-2 ) print('Looks ok!' ) print('Generating a caption...' ) UpperCAmelCase__ = '' UpperCAmelCase__ = tokenizer(lowerCamelCase , return_tensors='pt' ).input_ids.to(lowerCamelCase ) UpperCAmelCase__ = original_model.generate({'image': original_pixel_values} ) UpperCAmelCase__ = hf_model.generate( lowerCamelCase , lowerCamelCase , do_sample=lowerCamelCase , num_beams=5 , max_length=3_0 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print('Original generation:' , lowerCamelCase ) UpperCAmelCase__ = input_ids.shape[1] UpperCAmelCase__ = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=lowerCamelCase ) UpperCAmelCase__ = [text.strip() for text in output_text] print('HF generation:' , lowerCamelCase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(lowerCamelCase ) hf_model.save_pretrained(lowerCamelCase ) if push_to_hub: processor.push_to_hub(f'''nielsr/{model_name}''' ) hf_model.push_to_hub(f'''nielsr/{model_name}''' ) if __name__ == "__main__": lowerCAmelCase__ : Union[str, Any] = argparse.ArgumentParser() lowerCAmelCase__ : List[Any] = [ 'blip2-opt-2.7b', 'blip2-opt-6.7b', 'blip2-opt-2.7b-coco', 'blip2-opt-6.7b-coco', 'blip2-flan-t5-xl', 'blip2-flan-t5-xl-coco', 'blip2-flan-t5-xxl', ] parser.add_argument( '--model_name', default='blip2-opt-2.7b', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) lowerCAmelCase__ : List[Any] = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
704
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
632
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ["pixel_values"] def __init__( self : Union[str, Any] ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = PIL.Image.BICUBIC ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : Union[int, float] = 1 / 255 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,**lowerCamelCase__ : Any ,): super().__init__(**lowerCamelCase__ ) UpperCAmelCase__ = size if size is not None else {'height': 256, 'width': 256} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) UpperCAmelCase__ = crop_size if crop_size is not None else {'height': 224, 'width': 224} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='crop_size' ) UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = resample UpperCAmelCase__ = do_center_crop UpperCAmelCase__ = crop_size UpperCAmelCase__ = do_rescale UpperCAmelCase__ = rescale_factor UpperCAmelCase__ = do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : PILImageResampling = PIL.Image.BICUBIC ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : str ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return resize( lowerCamelCase__ ,size=(size['height'], size['width']) ,resample=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Dict ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return center_crop(lowerCamelCase__ ,size=(size['height'], size['width']) ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[int, float] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Tuple ,): return rescale(lowerCamelCase__ ,scale=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Union[str, Any] ,): return normalize(lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : ImageInput ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : float = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,lowerCamelCase__ : ChannelDimension = ChannelDimension.FIRST ,**lowerCamelCase__ : List[Any] ,): UpperCAmelCase__ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ = resample if resample is not None else self.resample UpperCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase__ = image_std if image_std is not None else self.image_std UpperCAmelCase__ = size if size is not None else self.size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) UpperCAmelCase__ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='crop_size' ) UpperCAmelCase__ = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. UpperCAmelCase__ = [to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: UpperCAmelCase__ = [self.resize(image=lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ) for image in images] if do_center_crop: UpperCAmelCase__ = [self.center_crop(image=lowerCamelCase__ ,size=lowerCamelCase__ ) for image in images] if do_rescale: UpperCAmelCase__ = [self.rescale(image=lowerCamelCase__ ,scale=lowerCamelCase__ ) for image in images] if do_normalize: UpperCAmelCase__ = [self.normalize(image=lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ) for image in images] UpperCAmelCase__ = [to_channel_dimension_format(lowerCamelCase__ ,lowerCamelCase__ ) for image in images] UpperCAmelCase__ = {'pixel_values': images} return BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ )
705
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
632
0
"""simple docstring""" import re def a_ ( lowerCamelCase ): if len(re.findall('[ATCG]' , lowerCamelCase ) ) != len(lowerCamelCase ): raise ValueError('Invalid Strand' ) return dna.translate(dna.maketrans('ATCG' , 'TAGC' ) ) if __name__ == "__main__": import doctest doctest.testmod()
706
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ : int = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Union[str, Any] = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
632
0
"""simple docstring""" from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case : """simple docstring""" def __init__( self : List[str] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Dict=3 ,lowerCamelCase__ : List[str]=32 ,lowerCamelCase__ : Tuple=3 ,lowerCamelCase__ : List[str]=10 ,lowerCamelCase__ : str=[10, 20, 30, 40] ,lowerCamelCase__ : List[str]=[1, 1, 2, 1] ,lowerCamelCase__ : List[str]=True ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : List[str]="relu" ,lowerCamelCase__ : Optional[Any]=3 ,lowerCamelCase__ : str=None ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = embeddings_size UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = depths UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = hidden_act UpperCAmelCase__ = num_labels UpperCAmelCase__ = scope UpperCAmelCase__ = len(lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] ,self.num_labels ) UpperCAmelCase__ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self : List[str] ): return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = TFRegNetModel(config=lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ,training=lowerCamelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Dict ,lowerCamelCase__ : int ): UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFRegNetForImageClassification(lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ,labels=lowerCamelCase__ ,training=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () snake_case__ = ( {"feature-extraction": TFRegNetModel, "image-classification": TFRegNetForImageClassification} if is_tf_available() else {} ) snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = TFRegNetModelTester(self ) UpperCAmelCase__ = ConfigTester(self ,config_class=lowerCamelCase__ ,has_text_modality=lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): return @unittest.skip(reason='RegNet does not use inputs_embeds' ) def __lowerCAmelCase ( self : Any ): pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 ,reason='TF does not support backprop for grouped convolutions on CPU.' ,) @slow def __lowerCAmelCase ( self : Optional[int] ): super().test_keras_fit() @unittest.skip(reason='RegNet does not support input and output embeddings' ) def __lowerCAmelCase ( self : Dict ): pass def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(lowerCamelCase__ ) UpperCAmelCase__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ['pixel_values'] self.assertListEqual(arg_names[:1] ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): def check_hidden_states_output(lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = model_class(lowerCamelCase__ ) UpperCAmelCase__ = model(**self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) ,training=lowerCamelCase__ ) UpperCAmelCase__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase__ = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) ,expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 2, self.model_tester.image_size // 2] ,) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase__ = layer_type UpperCAmelCase__ = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Any={} ): UpperCAmelCase__ = model(lowerCamelCase__ ,return_dict=lowerCamelCase__ ,**lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ,return_dict=lowerCamelCase__ ,**lowerCamelCase__ ).to_tuple() def recursive_check(lowerCamelCase__ : str ,lowerCamelCase__ : int ): if isinstance(lowerCamelCase__ ,(List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowerCamelCase__ ,lowerCamelCase__ ): recursive_check(lowerCamelCase__ ,lowerCamelCase__ ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(lowerCamelCase__ ,lowerCamelCase__ ) ) ,msg=( 'Tuple and dict output are not equal. Difference:' f''' {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}''' ) ,) recursive_check(lowerCamelCase__ ,lowerCamelCase__ ) for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) check_equivalence(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) check_equivalence(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) check_equivalence(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,{'output_hidden_states': True} ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) UpperCAmelCase__ = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ,return_labels=lowerCamelCase__ ) check_equivalence(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,{'output_hidden_states': True} ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def __lowerCAmelCase ( self : Optional[int] ): for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = TFRegNetModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def a_ ( ): UpperCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class snake_case ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self : Dict ): return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=lowerCamelCase__ ,return_tensors='tf' ) # forward pass UpperCAmelCase__ = model(**lowerCamelCase__ ,training=lowerCamelCase__ ) # verify the logits UpperCAmelCase__ = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase__ ) UpperCAmelCase__ = tf.constant([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] ,lowerCamelCase__ ,atol=1e-4 )
707
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase__ : Optional[int] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( image=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
632
0
"""simple docstring""" import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = np.full((len(lowerCamelCase ), sequence_length, 2) , lowerCamelCase ) else: UpperCAmelCase__ = np.full((len(lowerCamelCase ), sequence_length) , lowerCamelCase ) for i, tensor in enumerate(lowerCamelCase ): if padding_side == "right": if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = tensor[:sequence_length] else: UpperCAmelCase__ = tensor[:sequence_length] else: if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = tensor[:sequence_length] else: UpperCAmelCase__ = tensor[:sequence_length] return out_tensor.tolist() def a_ ( lowerCamelCase ): UpperCAmelCase__ = ord(lowerCamelCase ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True UpperCAmelCase__ = unicodedata.category(lowerCamelCase ) if cat.startswith('P' ): return True return False @dataclass class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = 42 snake_case__ = True snake_case__ = None snake_case__ = None snake_case__ = -1_00 snake_case__ = "pt" def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Optional[Any] ): import torch UpperCAmelCase__ = 'label' if 'label' in features[0].keys() else 'labels' UpperCAmelCase__ = [feature[label_name] for feature in features] if label_name in features[0].keys() else None UpperCAmelCase__ = self.tokenizer.pad( lowerCamelCase__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='pt' if labels is None else None ,) if labels is None: return batch UpperCAmelCase__ = torch.tensor(batch['entity_ids'] ).shape[1] UpperCAmelCase__ = self.tokenizer.padding_side if padding_side == "right": UpperCAmelCase__ = [ list(lowerCamelCase__ ) + [self.label_pad_token_id] * (sequence_length - len(lowerCamelCase__ )) for label in labels ] else: UpperCAmelCase__ = [ [self.label_pad_token_id] * (sequence_length - len(lowerCamelCase__ )) + list(lowerCamelCase__ ) for label in labels ] UpperCAmelCase__ = [feature['ner_tags'] for feature in features] UpperCAmelCase__ = padding_tensor(lowerCamelCase__ ,-1 ,lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = [feature['original_entity_spans'] for feature in features] UpperCAmelCase__ = padding_tensor(lowerCamelCase__ ,(-1, -1) ,lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = {k: torch.tensor(lowerCamelCase__ ,dtype=torch.intaa ) for k, v in batch.items()} return batch
708
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def a_ ( lowerCamelCase = 2_0_0_0_0_0_0 ): UpperCAmelCase__ = [0] UpperCAmelCase__ = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target UpperCAmelCase__ = 0 # the area corresponding to the grid that gives the product closest to target UpperCAmelCase__ = 0 # an estimate of b, using the quadratic formula UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the triangle number corresponding to b_floor UpperCAmelCase__ = 42 # the triangle number corresponding to b_ceil UpperCAmelCase__ = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): UpperCAmelCase__ = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 UpperCAmelCase__ = floor(lowerCamelCase ) UpperCAmelCase__ = ceil(lowerCamelCase ) UpperCAmelCase__ = triangle_numbers[b_floor] UpperCAmelCase__ = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_first_guess * triangle_a UpperCAmelCase__ = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_second_guess * triangle_a UpperCAmelCase__ = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
632
0
import itertools import math def a_ ( lowerCamelCase ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def a_ ( ): UpperCAmelCase__ = 2 while True: if is_prime(lowerCamelCase ): yield num num += 1 def a_ ( lowerCamelCase = 1_0_0_0_1 ): return next(itertools.islice(prime_generator() , nth - 1 , lowerCamelCase ) ) if __name__ == "__main__": print(F"""{solution() = }""")
709
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : Optional[Any] = ['model.decoder.embed_positions.weights'] def a_ ( lowerCamelCase ): if "emb" in name: UpperCAmelCase__ = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: UpperCAmelCase__ = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: UpperCAmelCase__ = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: UpperCAmelCase__ = name.replace('linear1' , 'fc1' ) if "linear2" in name: UpperCAmelCase__ = name.replace('linear2' , 'fc2' ) if "norm1" in name: UpperCAmelCase__ = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: UpperCAmelCase__ = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: UpperCAmelCase__ = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: UpperCAmelCase__ = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: UpperCAmelCase__ = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase__ = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = list(state_dict.keys() ) UpperCAmelCase__ = {} for key in keys: UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) UpperCAmelCase__ = rename_keys(lowerCamelCase ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase__ = val[:hidden_size, :] UpperCAmelCase__ = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase__ = val else: UpperCAmelCase__ = val return state_dict, enc_dec_proj_state_dict def a_ ( lowerCamelCase ): if checkpoint == "small": # default config values UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif checkpoint == "medium": UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 2_4 elif checkpoint == "large": UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 3_2 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) UpperCAmelCase__ = MusicgenDecoderConfig( hidden_size=lowerCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=lowerCamelCase , num_attention_heads=lowerCamelCase , ) return config @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="cpu" ): UpperCAmelCase__ = MusicGen.get_pretrained(lowerCamelCase , device=lowerCamelCase ) UpperCAmelCase__ = decoder_config_from_checkpoint(lowerCamelCase ) UpperCAmelCase__ = fairseq_model.lm.state_dict() UpperCAmelCase__ , UpperCAmelCase__ = rename_state_dict( lowerCamelCase , hidden_size=decoder_config.hidden_size ) UpperCAmelCase__ = TaEncoderModel.from_pretrained('t5-base' ) UpperCAmelCase__ = EncodecModel.from_pretrained('facebook/encodec_32khz' ) UpperCAmelCase__ = MusicgenForCausalLM(lowerCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase__ , UpperCAmelCase__ = decoder.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model UpperCAmelCase__ = MusicgenForConditionalGeneration(text_encoder=lowerCamelCase , audio_encoder=lowerCamelCase , decoder=lowerCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(lowerCamelCase ) # check we can do a forward pass UpperCAmelCase__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase__ = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits if logits.shape != (8, 1, 2_0_4_8): raise ValueError('Incorrect shape for logits' ) # now construct the processor UpperCAmelCase__ = AutoTokenizer.from_pretrained('t5-base' ) UpperCAmelCase__ = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) UpperCAmelCase__ = MusicgenProcessor(feature_extractor=lowerCamelCase , tokenizer=lowerCamelCase ) # set the appropriate bos/pad token ids UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 2_0_4_8 # set other default generation config params UpperCAmelCase__ = int(3_0 * audio_encoder.config.frame_rate ) UpperCAmelCase__ = True UpperCAmelCase__ = 3.0 if pytorch_dump_folder is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(lowerCamelCase ) processor.save_pretrained(lowerCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(lowerCamelCase ) processor.push_to_hub(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
632
0
"""simple docstring""" import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = DebertaTokenizer snake_case__ = True snake_case__ = DebertaTokenizerFast def __lowerCAmelCase ( self : Union[str, Any] ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '[UNK]', ] UpperCAmelCase__ = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) UpperCAmelCase__ = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] UpperCAmelCase__ = {'unk_token': '[UNK]'} UpperCAmelCase__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file ,'w' ,encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.merges_file ,'w' ,encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowerCAmelCase ( self : Tuple ,**lowerCamelCase__ : Optional[int] ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : Any ): UpperCAmelCase__ = 'lower newer' UpperCAmelCase__ = 'lower newer' return input_text, output_text def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_tokenizer() UpperCAmelCase__ = 'lower newer' UpperCAmelCase__ = ['l', 'o', 'w', 'er', '\u0120', 'n', 'e', 'w', 'er'] UpperCAmelCase__ = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokens + [tokenizer.unk_token] UpperCAmelCase__ = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_tokenizer() UpperCAmelCase__ = tokenizer('Hello' ,'World' ) UpperCAmelCase__ = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd['token_type_ids'] ,lowerCamelCase__ ) @slow def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.tokenizer_class.from_pretrained('microsoft/deberta-base' ) UpperCAmelCase__ = tokenizer.encode('sequence builders' ,add_special_tokens=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode('multi-sequence build' ,add_special_tokens=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode( 'sequence builders' ,add_special_tokens=lowerCamelCase__ ,add_prefix_space=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode( 'sequence builders' ,'multi-sequence build' ,add_special_tokens=lowerCamelCase__ ,add_prefix_space=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ,lowerCamelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: UpperCAmelCase__ = tokenizer_class.from_pretrained('microsoft/deberta-base' ) UpperCAmelCase__ = [ 'ALBERT: A Lite BERT for Self-supervised Learning of Language Representations', 'ALBERT incorporates two parameter reduction techniques', 'The first one is a factorized embedding parameterization. By decomposing the large vocabulary' ' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of' ' vocabulary embedding.', ] UpperCAmelCase__ = tokenizer(lowerCamelCase__ ,padding=lowerCamelCase__ ) UpperCAmelCase__ = [tokenizer.decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ ) for seq in encoding['input_ids']] # fmt: off UpperCAmelCase__ = { 'input_ids': [ [1, 2_118, 11_126, 565, 35, 83, 25_191, 163, 18_854, 13, 12_156, 12, 16_101, 25_376, 13_807, 9, 22_205, 27_893, 1_635, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 2_118, 11_126, 565, 24_536, 80, 43_797, 4_878, 7_373, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 133, 78, 65, 16, 10, 3_724, 1_538, 33_183, 11_303, 43_797, 1_938, 4, 870, 24_165, 29_105, 5, 739, 32_644, 33_183, 11_303, 36_173, 88, 80, 650, 7_821, 45_940, 6, 52, 2_559, 5, 1_836, 9, 5, 7_397, 13_171, 31, 5, 1_836, 9, 32_644, 33_183, 11_303, 4, 2] ], 'token_type_ids': [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], 'attention_mask': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on UpperCAmelCase__ = [ 'ALBERT: A Lite BERT for Self-supervised Learning of Language Representations', 'ALBERT incorporates two parameter reduction techniques', 'The first one is a factorized embedding parameterization. By decomposing the large vocabulary' ' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of' ' vocabulary embedding.', ] self.assertDictEqual(encoding.data ,lowerCamelCase__ ) for expected, decoded in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ )
710
"""simple docstring""" lowerCAmelCase__ : Tuple = range(2, 20 + 1) lowerCAmelCase__ : Optional[Any] = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase__ : dict[int, dict[int, list[list[int]]]] = {} def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(a_i[j] for j in range(lowerCamelCase , len(lowerCamelCase ) ) ) UpperCAmelCase__ = sum(a_i[j] * base[j] for j in range(min(len(lowerCamelCase ) , lowerCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ = 0, 0 UpperCAmelCase__ = n - i UpperCAmelCase__ = memo.get(lowerCamelCase ) if sub_memo is not None: UpperCAmelCase__ = sub_memo.get(lowerCamelCase ) if jumps is not None and len(lowerCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ = -1 for _k in range(len(lowerCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ = diff + c for j in range(min(lowerCamelCase , len(lowerCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) if new_c > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: UpperCAmelCase__ = [] else: UpperCAmelCase__ = {c: []} UpperCAmelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , k - 1 , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ = compute(lowerCamelCase , lowerCamelCase , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ = 0 while j < len(lowerCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowerCamelCase , (diff, dn, k) ) return (diff, dn) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if i >= n: return 0, i if k > len(lowerCamelCase ): a_i.extend([0 for _ in range(k - len(lowerCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 0, 0, 0 for j in range(len(lowerCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ = ds_c + ds_b diff += addend UpperCAmelCase__ = 0 for j in range(lowerCamelCase ): UpperCAmelCase__ = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return diff, i - start_i def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): for j in range(lowerCamelCase , len(lowerCamelCase ) ): UpperCAmelCase__ = digits[j] + addend if s >= 1_0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) UpperCAmelCase__ = addend // 1_0 + quotient else: UpperCAmelCase__ = s UpperCAmelCase__ = addend // 1_0 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) digits.append(lowerCamelCase ) def a_ ( lowerCamelCase = 1_0**1_5 ): UpperCAmelCase__ = [1] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , 2_0 , i + dn , lowerCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ = 0 for j in range(len(lowerCamelCase ) ): a_n += digits[j] * 1_0**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): if isinstance(lowerCamelCase , lowerCamelCase ) and isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = len(set_a.intersection(lowerCamelCase ) ) if alternative_union: UpperCAmelCase__ = len(lowerCamelCase ) + len(lowerCamelCase ) else: UpperCAmelCase__ = len(set_a.union(lowerCamelCase ) ) return intersection / union if isinstance(lowerCamelCase , (list, tuple) ) and isinstance(lowerCamelCase , (list, tuple) ): UpperCAmelCase__ = [element for element in set_a if element in set_b] if alternative_union: UpperCAmelCase__ = len(lowerCamelCase ) + len(lowerCamelCase ) return len(lowerCamelCase ) / union else: UpperCAmelCase__ = set_a + [element for element in set_b if element not in set_a] return len(lowerCamelCase ) / len(lowerCamelCase ) return len(lowerCamelCase ) / len(lowerCamelCase ) return None if __name__ == "__main__": lowerCAmelCase__ : Union[str, Any] = {'a', 'b', 'c', 'd', 'e'} lowerCAmelCase__ : List[str] = {'c', 'd', 'e', 'f', 'h', 'i'} print(jaccard_similarity(set_a, set_b))
711
"""simple docstring""" import random class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : str ): UpperCAmelCase__ = [ord(lowerCamelCase__ ) for i in text] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i in plain: UpperCAmelCase__ = random.randint(1 ,300 ) UpperCAmelCase__ = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): UpperCAmelCase__ = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ : Dict = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
632
0
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def a_ ( lowerCamelCase ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def a_ ( lowerCamelCase ): # word like '180' or '身高' or '神' for char in word: UpperCAmelCase__ = ord(lowerCamelCase ) if not _is_chinese_char(lowerCamelCase ): return 0 return 1 def a_ ( lowerCamelCase ): UpperCAmelCase__ = set() for token in tokens: UpperCAmelCase__ = len(lowerCamelCase ) > 1 and is_chinese(lowerCamelCase ) if chinese_word: word_set.add(lowerCamelCase ) UpperCAmelCase__ = list(lowerCamelCase ) return word_list def a_ ( lowerCamelCase , lowerCamelCase ): if not chinese_word_set: return bert_tokens UpperCAmelCase__ = max([len(lowerCamelCase ) for w in chinese_word_set] ) UpperCAmelCase__ = bert_tokens UpperCAmelCase__ , UpperCAmelCase__ = 0, len(lowerCamelCase ) while start < end: UpperCAmelCase__ = True if is_chinese(bert_word[start] ): UpperCAmelCase__ = min(end - start , lowerCamelCase ) for i in range(lowerCamelCase , 1 , -1 ): UpperCAmelCase__ = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): UpperCAmelCase__ = '##' + bert_word[j] UpperCAmelCase__ = start + i UpperCAmelCase__ = False break if single_word: start += 1 return bert_word def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = [] for i in range(0 , len(lowerCamelCase ) , 1_0_0 ): UpperCAmelCase__ = ltp_tokenizer.pipeline(lines[i : i + 1_0_0] , tasks=['cws'] ).cws UpperCAmelCase__ = [get_chinese_word(lowerCamelCase ) for r in res] ltp_res.extend(lowerCamelCase ) assert len(lowerCamelCase ) == len(lowerCamelCase ) UpperCAmelCase__ = [] for i in range(0 , len(lowerCamelCase ) , 1_0_0 ): UpperCAmelCase__ = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=lowerCamelCase , truncation=lowerCamelCase , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(lowerCamelCase ) == len(lowerCamelCase ) UpperCAmelCase__ = [] for input_ids, chinese_word in zip(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = [] for id in input_ids: UpperCAmelCase__ = bert_tokenizer._convert_id_to_token(lowerCamelCase ) input_tokens.append(lowerCamelCase ) UpperCAmelCase__ = add_sub_symbol(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(lowerCamelCase ): if token[:2] == "##": UpperCAmelCase__ = token[2:] # save chinese tokens' pos if len(lowerCamelCase ) == 1 and _is_chinese_char(ord(lowerCamelCase ) ): ref_id.append(lowerCamelCase ) ref_ids.append(lowerCamelCase ) assert len(lowerCamelCase ) == len(lowerCamelCase ) return ref_ids def a_ ( lowerCamelCase ): # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name , 'r' , encoding='utf-8' ) as f: UpperCAmelCase__ = f.readlines() UpperCAmelCase__ = [line.strip() for line in data if len(lowerCamelCase ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' UpperCAmelCase__ = LTP(args.ltp ) # faster in GPU device UpperCAmelCase__ = BertTokenizer.from_pretrained(args.bert ) UpperCAmelCase__ = prepare_ref(lowerCamelCase , lowerCamelCase , lowerCamelCase ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: UpperCAmelCase__ = [json.dumps(lowerCamelCase ) + '\n' for ref in ref_ids] f.writelines(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : List[str] = argparse.ArgumentParser(description='prepare_chinese_ref') parser.add_argument( '--file_name', required=False, type=str, default='./resources/chinese-demo.txt', help='file need process, same as training data in lm', ) parser.add_argument( '--ltp', required=False, type=str, default='./resources/ltp', help='resources for LTP tokenizer, usually a path', ) parser.add_argument( '--bert', required=False, type=str, default='./resources/robert', help='resources for Bert tokenizer', ) parser.add_argument( '--save_path', required=False, type=str, default='./resources/ref.txt', help='path to save res', ) lowerCAmelCase__ : Optional[int] = parser.parse_args() main(args)
712
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
632
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ : List[Any] = {'processing_layoutxlm': ['LayoutXLMProcessor']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Dict = ['LayoutXLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[str] = ['LayoutXLMTokenizerFast'] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys lowerCAmelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
713
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Dict ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_torch def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,torch.tensor(lowerCamelCase__ ) ,torch.tensor(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) @require_vision @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_tf def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,tf.convert_to_tensor(lowerCamelCase__ ) ,tf.convert_to_tensor(lowerCamelCase__ ) ,return_tensors='tf' ,) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowerCamelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = np.random.randint(0 ,2 ,size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase__ = [tf.convert_to_tensor(lowerCamelCase__ )] UpperCAmelCase__ = [torch.tensor(lowerCamelCase__ )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='pt' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) )
632
0
"""simple docstring""" def a_ ( lowerCamelCase = 2_0_0 ): UpperCAmelCase__ = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] UpperCAmelCase__ = [0] * (pence + 1) UpperCAmelCase__ = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowerCamelCase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73_682
714
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) lowerCAmelCase__ : str = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "ctrl" snake_case__ = ["past_key_values"] snake_case__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any ,lowerCamelCase__ : str=246_534 ,lowerCamelCase__ : List[str]=256 ,lowerCamelCase__ : Optional[int]=1_280 ,lowerCamelCase__ : Any=8_192 ,lowerCamelCase__ : int=48 ,lowerCamelCase__ : Optional[Any]=16 ,lowerCamelCase__ : Union[str, Any]=0.1 ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : List[str]=1e-6 ,lowerCamelCase__ : List[str]=0.0_2 ,lowerCamelCase__ : Tuple=True ,**lowerCamelCase__ : Optional[Any] ,): UpperCAmelCase__ = vocab_size UpperCAmelCase__ = n_positions UpperCAmelCase__ = n_embd UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = dff UpperCAmelCase__ = resid_pdrop UpperCAmelCase__ = embd_pdrop UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache super().__init__(**lowerCamelCase__ )
632
0
"""simple docstring""" import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = StableDiffusionDiffEditPipeline snake_case__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"height", "width", "image"} | {"image_latents"} snake_case__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"image"} | {"image_latents"} snake_case__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess snake_case__ = frozenset([] ) def __lowerCAmelCase ( self : Optional[int] ): torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') ,up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') ,cross_attention_dim=32 ,attention_head_dim=(2, 4) ,use_linear_projection=lowerCamelCase__ ,) UpperCAmelCase__ = DDIMScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule='scaled_linear' ,clip_sample=lowerCamelCase__ ,set_alpha_to_one=lowerCamelCase__ ,) UpperCAmelCase__ = DDIMInverseScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule='scaled_linear' ,clip_sample=lowerCamelCase__ ,set_alpha_to_zero=lowerCamelCase__ ,) torch.manual_seed(0 ) UpperCAmelCase__ = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=4 ,sample_size=128 ,) torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,hidden_act='gelu' ,projection_dim=512 ,) UpperCAmelCase__ = CLIPTextModel(lowerCamelCase__ ) UpperCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) UpperCAmelCase__ = { 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : int=0 ): UpperCAmelCase__ = floats_tensor((1, 16, 16) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = floats_tensor((1, 2, 4, 16, 16) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Dict ,lowerCamelCase__ : int=0 ): UpperCAmelCase__ = floats_tensor((1, 3, 32, 32) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = image.cpu().permute(0 ,2 ,3 ,1 )[0] UpperCAmelCase__ = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : str=0 ): UpperCAmelCase__ = floats_tensor((1, 3, 32, 32) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) UpperCAmelCase__ = image.cpu().permute(0 ,2 ,3 ,1 )[0] UpperCAmelCase__ = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : Any ): if not hasattr(self.pipeline_class ,'_optional_components' ): return UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) UpperCAmelCase__ = self.get_dummy_inputs(lowerCamelCase__ ) UpperCAmelCase__ = pipe(**lowerCamelCase__ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = self.pipeline_class.from_pretrained(lowerCamelCase__ ) pipe_loaded.to(lowerCamelCase__ ) pipe_loaded.set_progress_bar_config(disable=lowerCamelCase__ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCamelCase__ ,lowerCamelCase__ ) is None ,f'''`{optional_component}` did not stay set to None after loading.''' ,) UpperCAmelCase__ = self.get_dummy_inputs(lowerCamelCase__ ) UpperCAmelCase__ = pipe_loaded(**lowerCamelCase__ )[0] UpperCAmelCase__ = np.abs(output - output_loaded ).max() self.assertLess(lowerCamelCase__ ,1e-4 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = 'cpu' UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = self.get_dummy_mask_inputs(lowerCamelCase__ ) UpperCAmelCase__ = pipe.generate_mask(**lowerCamelCase__ ) UpperCAmelCase__ = mask[0, -3:, -3:] self.assertEqual(mask.shape ,(1, 16, 16) ) UpperCAmelCase__ = np.array([0] * 9 ) UpperCAmelCase__ = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ ,1e-3 ) self.assertEqual(mask[0, -3, -4] ,0 ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = 'cpu' UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = self.get_dummy_inversion_inputs(lowerCamelCase__ ) UpperCAmelCase__ = pipe.invert(**lowerCamelCase__ ).images UpperCAmelCase__ = image[0, -1, -3:, -3:] self.assertEqual(image.shape ,(2, 32, 32, 3) ) UpperCAmelCase__ = np.array( [0.5_1_5_0, 0.5_1_3_4, 0.5_0_4_3, 0.5_3_7_6, 0.4_6_9_4, 0.5_1_0_5_0, 0.5_0_1_5, 0.4_4_0_7, 0.4_7_9_9] ,) UpperCAmelCase__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ ,1e-3 ) def __lowerCAmelCase ( self : Optional[int] ): super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = 'cpu' UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = {'beta_start': 0.0_0_0_8_5, 'beta_end': 0.0_1_2, 'beta_schedule': 'scaled_linear'} UpperCAmelCase__ = DPMSolverMultistepScheduler(**lowerCamelCase__ ) UpperCAmelCase__ = DPMSolverMultistepInverseScheduler(**lowerCamelCase__ ) UpperCAmelCase__ = self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = self.get_dummy_inversion_inputs(lowerCamelCase__ ) UpperCAmelCase__ = pipe.invert(**lowerCamelCase__ ).images UpperCAmelCase__ = image[0, -1, -3:, -3:] self.assertEqual(image.shape ,(2, 32, 32, 3) ) UpperCAmelCase__ = np.array( [0.5_1_5_0, 0.5_1_3_4, 0.5_0_4_3, 0.5_3_7_6, 0.4_6_9_4, 0.5_1_0_5_0, 0.5_0_1_5, 0.4_4_0_7, 0.4_7_9_9] ,) UpperCAmelCase__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ ,1e-3 ) @require_torch_gpu @slow class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __lowerCAmelCase ( cls : Dict ): UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) UpperCAmelCase__ = raw_image.convert('RGB' ).resize((768, 768) ) UpperCAmelCase__ = raw_image def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' ,safety_checker=lowerCamelCase__ ,torch_dtype=torch.floataa ) UpperCAmelCase__ = DDIMScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase__ = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'a bowl of fruit' UpperCAmelCase__ = 'a bowl of pears' UpperCAmelCase__ = pipe.generate_mask( image=self.raw_image ,source_prompt=lowerCamelCase__ ,target_prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,) UpperCAmelCase__ = pipe.invert( prompt=lowerCamelCase__ ,image=self.raw_image ,inpaint_strength=0.7 ,generator=lowerCamelCase__ ).latents UpperCAmelCase__ = pipe( prompt=lowerCamelCase__ ,mask_image=lowerCamelCase__ ,image_latents=lowerCamelCase__ ,generator=lowerCamelCase__ ,negative_prompt=lowerCamelCase__ ,inpaint_strength=0.7 ,output_type='numpy' ,).images[0] UpperCAmelCase__ = ( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1 def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' ,safety_checker=lowerCamelCase__ ,torch_dtype=torch.floataa ) UpperCAmelCase__ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase__ = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'a bowl of fruit' UpperCAmelCase__ = 'a bowl of pears' UpperCAmelCase__ = pipe.generate_mask( image=self.raw_image ,source_prompt=lowerCamelCase__ ,target_prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,) UpperCAmelCase__ = pipe.invert( prompt=lowerCamelCase__ ,image=self.raw_image ,inpaint_strength=0.7 ,generator=lowerCamelCase__ ,num_inference_steps=25 ,).latents UpperCAmelCase__ = pipe( prompt=lowerCamelCase__ ,mask_image=lowerCamelCase__ ,image_latents=lowerCamelCase__ ,generator=lowerCamelCase__ ,negative_prompt=lowerCamelCase__ ,inpaint_strength=0.7 ,num_inference_steps=25 ,output_type='numpy' ,).images[0] UpperCAmelCase__ = ( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1
715
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCAmelCase__ : Dict = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def a_ ( ): UpperCAmelCase__ = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCAmelCase__ = get_sagemaker_input() else: UpperCAmelCase__ = get_cluster_input() return config def a_ ( lowerCamelCase=None ): if subparsers is not None: UpperCAmelCase__ = subparsers.add_parser('config' , description=lowerCamelCase ) else: UpperCAmelCase__ = argparse.ArgumentParser('Accelerate config command' , description=lowerCamelCase ) parser.add_argument( '--config_file' , default=lowerCamelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def a_ ( lowerCamelCase ): UpperCAmelCase__ = get_user_input() if args.config_file is not None: UpperCAmelCase__ = args.config_file else: if not os.path.isdir(lowerCamelCase ): os.makedirs(lowerCamelCase ) UpperCAmelCase__ = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(lowerCamelCase ) else: config.to_yaml_file(lowerCamelCase ) print(f'''accelerate configuration saved at {config_file}''' ) def a_ ( ): UpperCAmelCase__ = config_command_parser() UpperCAmelCase__ = parser.parse_args() config_command(lowerCamelCase ) if __name__ == "__main__": main()
632
0
"""simple docstring""" import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging lowerCAmelCase__ : str = ( 'https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py' ) lowerCAmelCase__ : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name def a_ ( ): UpperCAmelCase__ = 'https://pypi.org/pypi/diffusers/json' UpperCAmelCase__ = json.loads(request.urlopen(lowerCamelCase ).read() )['releases'].keys() return sorted(lowerCamelCase , key=lambda lowerCamelCase : version.Version(lowerCamelCase ) ) def a_ ( ): # This function has already been executed if HF_MODULES_CACHE already is in the Python path. if HF_MODULES_CACHE in sys.path: return sys.path.append(lowerCamelCase ) os.makedirs(lowerCamelCase , exist_ok=lowerCamelCase ) UpperCAmelCase__ = Path(lowerCamelCase ) / '__init__.py' if not init_path.exists(): init_path.touch() def a_ ( lowerCamelCase ): init_hf_modules() UpperCAmelCase__ = Path(lowerCamelCase ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(lowerCamelCase , exist_ok=lowerCamelCase ) UpperCAmelCase__ = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def a_ ( lowerCamelCase ): with open(lowerCamelCase , 'r' , encoding='utf-8' ) as f: UpperCAmelCase__ = f.read() # Imports of the form `import .xxx` UpperCAmelCase__ = re.findall('^\s*import\s+\.(\S+)\s*$' , lowerCamelCase , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , lowerCamelCase , flags=re.MULTILINE ) # Unique-ify return list(set(lowerCamelCase ) ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = False UpperCAmelCase__ = [module_file] UpperCAmelCase__ = [] # Let's recurse through all relative imports while not no_change: UpperCAmelCase__ = [] for f in files_to_check: new_imports.extend(get_relative_imports(lowerCamelCase ) ) UpperCAmelCase__ = Path(lowerCamelCase ).parent UpperCAmelCase__ = [str(module_path / m ) for m in new_imports] UpperCAmelCase__ = [f for f in new_import_files if f not in all_relative_imports] UpperCAmelCase__ = [f'''{f}.py''' for f in new_import_files] UpperCAmelCase__ = len(lowerCamelCase ) == 0 all_relative_imports.extend(lowerCamelCase ) return all_relative_imports def a_ ( lowerCamelCase ): with open(lowerCamelCase , 'r' , encoding='utf-8' ) as f: UpperCAmelCase__ = f.read() # Imports of the form `import xxx` UpperCAmelCase__ = re.findall('^\s*import\s+(\S+)\s*$' , lowerCamelCase , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , lowerCamelCase , flags=re.MULTILINE ) # Only keep the top-level module UpperCAmelCase__ = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all UpperCAmelCase__ = list(set(lowerCamelCase ) ) UpperCAmelCase__ = [] for imp in imports: try: importlib.import_module(lowerCamelCase ) except ImportError: missing_packages.append(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f'''{", ".join(lowerCamelCase )}. Run `pip install {" ".join(lowerCamelCase )}`''' ) return get_relative_imports(lowerCamelCase ) def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = module_path.replace(os.path.sep , '.' ) UpperCAmelCase__ = importlib.import_module(lowerCamelCase ) if class_name is None: return find_pipeline_class(lowerCamelCase ) return getattr(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase ): from ..pipelines import DiffusionPipeline UpperCAmelCase__ = dict(inspect.getmembers(lowerCamelCase , inspect.isclass ) ) UpperCAmelCase__ = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , lowerCamelCase ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f'''Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:''' f''' {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in''' f''' {loaded_module}.''' ) UpperCAmelCase__ = cls return pipeline_class def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , ): UpperCAmelCase__ = str(lowerCamelCase ) UpperCAmelCase__ = os.path.join(lowerCamelCase , lowerCamelCase ) if os.path.isfile(lowerCamelCase ): UpperCAmelCase__ = module_file_or_url UpperCAmelCase__ = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: UpperCAmelCase__ = get_diffusers_versions() # cut ".dev0" UpperCAmelCase__ = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: UpperCAmelCase__ = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f'''Defaulting to latest_version: {revision}.''' ) elif revision in available_versions: UpperCAmelCase__ = f'''v{revision}''' elif revision == "main": UpperCAmelCase__ = revision else: raise ValueError( f'''`custom_revision`: {revision} does not exist. Please make sure to choose one of''' f''' {", ".join(available_versions + ["main"] )}.''' ) # community pipeline on GitHub UpperCAmelCase__ = COMMUNITY_PIPELINES_URL.format(revision=lowerCamelCase , pipeline=lowerCamelCase ) try: UpperCAmelCase__ = cached_download( lowerCamelCase , cache_dir=lowerCamelCase , force_download=lowerCamelCase , proxies=lowerCamelCase , resume_download=lowerCamelCase , local_files_only=lowerCamelCase , use_auth_token=lowerCamelCase , ) UpperCAmelCase__ = 'git' UpperCAmelCase__ = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f'''Could not locate the {module_file} inside {pretrained_model_name_or_path}.''' ) raise else: try: # Load from URL or cache if already cached UpperCAmelCase__ = hf_hub_download( lowerCamelCase , lowerCamelCase , cache_dir=lowerCamelCase , force_download=lowerCamelCase , proxies=lowerCamelCase , resume_download=lowerCamelCase , local_files_only=lowerCamelCase , use_auth_token=lowerCamelCase , ) UpperCAmelCase__ = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f'''Could not locate the {module_file} inside {pretrained_model_name_or_path}.''' ) raise # Check we have all the requirements in our environment UpperCAmelCase__ = check_imports(lowerCamelCase ) # Now we move the module inside our cached dynamic modules. UpperCAmelCase__ = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(lowerCamelCase ) UpperCAmelCase__ = Path(lowerCamelCase ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(lowerCamelCase , submodule_path / module_file ) for module_needed in modules_needed: UpperCAmelCase__ = f'''{module_needed}.py''' shutil.copy(os.path.join(lowerCamelCase , lowerCamelCase ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = use_auth_token elif use_auth_token is True: UpperCAmelCase__ = HfFolder.get_token() else: UpperCAmelCase__ = None UpperCAmelCase__ = model_info(lowerCamelCase , revision=lowerCamelCase , token=lowerCamelCase ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. UpperCAmelCase__ = submodule_path / commit_hash UpperCAmelCase__ = full_submodule + os.path.sep + commit_hash create_dynamic_module(lowerCamelCase ) if not (submodule_path / module_file).exists(): shutil.copy(lowerCamelCase , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( lowerCamelCase , f'''{module_needed}.py''' , cache_dir=lowerCamelCase , force_download=lowerCamelCase , resume_download=lowerCamelCase , proxies=lowerCamelCase , use_auth_token=lowerCamelCase , revision=lowerCamelCase , local_files_only=lowerCamelCase , ) return os.path.join(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , **lowerCamelCase , ): UpperCAmelCase__ = get_cached_module_file( lowerCamelCase , lowerCamelCase , cache_dir=lowerCamelCase , force_download=lowerCamelCase , resume_download=lowerCamelCase , proxies=lowerCamelCase , use_auth_token=lowerCamelCase , revision=lowerCamelCase , local_files_only=lowerCamelCase , ) return get_class_in_module(lowerCamelCase , final_module.replace('.py' , '' ) )
716
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): return x if y == 0 else greatest_common_divisor(lowerCamelCase , x % y ) def a_ ( lowerCamelCase , lowerCamelCase ): return (x * y) // greatest_common_divisor(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase = 2_0 ): UpperCAmelCase__ = 1 for i in range(1 , n + 1 ): UpperCAmelCase__ = lcm(lowerCamelCase , lowerCamelCase ) return g if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase__ : List[Any] = { 'google/efficientnet-b7': 'https://huggingface.co/google/efficientnet-b7/resolve/main/config.json', } class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "efficientnet" def __init__( self : List[str] ,lowerCamelCase__ : int = 3 ,lowerCamelCase__ : int = 600 ,lowerCamelCase__ : float = 2.0 ,lowerCamelCase__ : float = 3.1 ,lowerCamelCase__ : int = 8 ,lowerCamelCase__ : List[int] = [3, 3, 5, 3, 5, 5, 3] ,lowerCamelCase__ : List[int] = [32, 16, 24, 40, 80, 112, 192] ,lowerCamelCase__ : List[int] = [16, 24, 40, 80, 112, 192, 320] ,lowerCamelCase__ : List[int] = [] ,lowerCamelCase__ : List[int] = [1, 2, 2, 2, 1, 2, 1] ,lowerCamelCase__ : List[int] = [1, 2, 2, 3, 3, 4, 1] ,lowerCamelCase__ : List[int] = [1, 6, 6, 6, 6, 6, 6] ,lowerCamelCase__ : float = 0.2_5 ,lowerCamelCase__ : str = "swish" ,lowerCamelCase__ : int = 2_560 ,lowerCamelCase__ : str = "mean" ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : float = 0.0_0_1 ,lowerCamelCase__ : float = 0.9_9 ,lowerCamelCase__ : float = 0.5 ,lowerCamelCase__ : float = 0.2 ,**lowerCamelCase__ : Dict ,): super().__init__(**lowerCamelCase__ ) UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = width_coefficient UpperCAmelCase__ = depth_coefficient UpperCAmelCase__ = depth_divisor UpperCAmelCase__ = kernel_sizes UpperCAmelCase__ = in_channels UpperCAmelCase__ = out_channels UpperCAmelCase__ = depthwise_padding UpperCAmelCase__ = strides UpperCAmelCase__ = num_block_repeats UpperCAmelCase__ = expand_ratios UpperCAmelCase__ = squeeze_expansion_ratio UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dim UpperCAmelCase__ = pooling_type UpperCAmelCase__ = initializer_range UpperCAmelCase__ = batch_norm_eps UpperCAmelCase__ = batch_norm_momentum UpperCAmelCase__ = dropout_rate UpperCAmelCase__ = drop_connect_rate UpperCAmelCase__ = sum(lowerCamelCase__ ) * 4 class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = version.parse("1.11" ) @property def __lowerCAmelCase ( self : int ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def __lowerCAmelCase ( self : Optional[Any] ): return 1e-5
717
"""simple docstring""" import warnings from functools import wraps from typing import Callable def a_ ( lowerCamelCase ): @wraps(lowerCamelCase ) def _inner_fn(*lowerCamelCase , **lowerCamelCase ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , lowerCamelCase , ) return fn(*lowerCamelCase , **lowerCamelCase ) return _inner_fn
632
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline lowerCAmelCase__ : int = logging.get_logger(__name__) # pylint: disable=invalid-name class snake_case ( __UpperCAmelCase ): """simple docstring""" def __init__( self : List[str] ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): super().__init__() self.register_modules(unet=lowerCamelCase__ ,scheduler=lowerCamelCase__ ) @torch.no_grad() def __call__( self : List[str] ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : int = 100 ,lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,lowerCamelCase__ : Optional[float] = None ,lowerCamelCase__ : bool = True ,): if audio_length_in_s is None: UpperCAmelCase__ = self.unet.config.sample_size / self.unet.config.sample_rate UpperCAmelCase__ = audio_length_in_s * self.unet.config.sample_rate UpperCAmelCase__ = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( f'''{audio_length_in_s} is too small. Make sure it\'s bigger or equal to''' f''' {3 * down_scale_factor / self.unet.config.sample_rate}.''' ) UpperCAmelCase__ = int(lowerCamelCase__ ) if sample_size % down_scale_factor != 0: UpperCAmelCase__ = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( f'''{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled''' f''' by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising''' ' process.' ) UpperCAmelCase__ = int(lowerCamelCase__ ) UpperCAmelCase__ = next(iter(self.unet.parameters() ) ).dtype UpperCAmelCase__ = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) UpperCAmelCase__ = randn_tensor(lowerCamelCase__ ,generator=lowerCamelCase__ ,device=self.device ,dtype=lowerCamelCase__ ) # set step values self.scheduler.set_timesteps(lowerCamelCase__ ,device=audio.device ) UpperCAmelCase__ = self.scheduler.timesteps.to(lowerCamelCase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCAmelCase__ = self.unet(lowerCamelCase__ ,lowerCamelCase__ ).sample # 2. compute previous image: x_t -> t_t-1 UpperCAmelCase__ = self.scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample UpperCAmelCase__ = audio.clamp(-1 ,1 ).float().cpu().numpy() UpperCAmelCase__ = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=lowerCamelCase__ )
718
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCAmelCase__ : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCAmelCase__ : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowerCAmelCase__ : set[int] = {ord(char) for char in VALID_CHARS} lowerCAmelCase__ : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = "" UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 for keychar, cipherchar in zip(cycle(lowerCamelCase ) , lowerCamelCase ): UpperCAmelCase__ = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCamelCase ) return decoded def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] for key in product(lowerCamelCase , repeat=3 ): UpperCAmelCase__ = try_key(lowerCamelCase , lowerCamelCase ) if encoded is not None: possibles.append(lowerCamelCase ) return possibles def a_ ( lowerCamelCase , lowerCamelCase ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCamelCase = "p059_cipher.txt" ): UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = Path(lowerCamelCase ).parent.joinpath(lowerCamelCase ).read_text(encoding='utf-8' ) UpperCAmelCase__ = [int(lowerCamelCase ) for number in data.strip().split(',' )] UpperCAmelCase__ = filter_valid_chars(lowerCamelCase ) for common_word in COMMON_WORDS: UpperCAmelCase__ = filter_common_word(lowerCamelCase , lowerCamelCase ) if len(lowerCamelCase ) == 1: break UpperCAmelCase__ = possibles[0] return sum(ord(lowerCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
632
0
from manim import * class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = Rectangle(height=0.5 ,width=0.5 ) UpperCAmelCase__ = Rectangle(height=0.2_5 ,width=0.2_5 ) UpperCAmelCase__ = Rectangle(height=0.4_6 ,width=0.4_6 ).set_stroke(width=0 ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = VGroup(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = Text('CPU' ,font_size=24 ) UpperCAmelCase__ = Group(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0.5 ,aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) UpperCAmelCase__ = [mem.copy() for i in range(4 )] UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = Text('GPU' ,font_size=24 ) UpperCAmelCase__ = Group(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0.5 ,aligned_edge=lowerCamelCase__ ) gpu.move_to([-1, -1, 0] ) self.add(lowerCamelCase__ ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = Text('Model' ,font_size=24 ) UpperCAmelCase__ = Group(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0.5 ,aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.add(lowerCamelCase__ ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(lowerCamelCase__ ): rect.set_stroke(lowerCamelCase__ ) UpperCAmelCase__ = Rectangle(height=0.4_6 / 4 ,width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ ,opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) ,buff=0.0_2 ,direction=lowerCamelCase__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] ,direction=lowerCamelCase__ ,buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] ,direction=lowerCamelCase__ ,buff=0.0 ) self.add(lowerCamelCase__ ) model_cpu_arr.append(lowerCamelCase__ ) self.add(*lowerCamelCase__ ,*lowerCamelCase__ ,*lowerCamelCase__ ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = Text('Loaded Checkpoint' ,font_size=24 ) UpperCAmelCase__ = Group(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0.5 ,aligned_edge=lowerCamelCase__ ) checkpoint.move_to([3, 0.5, 0] ) self.add(lowerCamelCase__ ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(lowerCamelCase__ ): UpperCAmelCase__ = fill.copy().set_fill(lowerCamelCase__ ,opacity=0.7 ) target.move_to(lowerCamelCase__ ) ckpt_arr.append(lowerCamelCase__ ) UpperCAmelCase__ = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(lowerCamelCase__ ) self.add(*lowerCamelCase__ ,*lowerCamelCase__ ) UpperCAmelCase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ = MarkupText( f'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' ,font_size=18 ,) key_text.move_to([-5, 2.4, 0] ) self.add(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = MarkupText( f'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' ,font_size=18 ,) blue_text.next_to(lowerCamelCase__ ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) self.add(lowerCamelCase__ ) UpperCAmelCase__ = MarkupText( f'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''' ,font_size=24 ,) step_a.move_to([2, 2, 0] ) UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = VGroup(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0 ) UpperCAmelCase__ = Text('Disk' ,font_size=24 ) UpperCAmelCase__ = Group(lowerCamelCase__ ,lowerCamelCase__ ).arrange(lowerCamelCase__ ,buff=0.5 ,aligned_edge=lowerCamelCase__ ) disk.move_to([-4.0, -1.2_5, 0] ) self.play(Write(lowerCamelCase__ ,run_time=3 ) ,Write(lowerCamelCase__ ,run_time=1 ) ,Create(lowerCamelCase__ ,run_time=1 ) ) UpperCAmelCase__ = [] for i, rect in enumerate(lowerCamelCase__ ): UpperCAmelCase__ = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(lowerCamelCase__ ,run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(FadeOut(lowerCamelCase__ ) ) UpperCAmelCase__ = MarkupText(f'''Then, the checkpoint is removed from memory\nthrough garbage collection.''' ,font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ ,run_time=3 ) ) self.play( FadeOut(lowerCamelCase__ ,lowerCamelCase__ ,*lowerCamelCase__ ,*lowerCamelCase__ ) ,) self.wait()
719
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = XLMProphetNetTokenizer snake_case__ = False snake_case__ = True def __lowerCAmelCase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = '[PAD]' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'[PAD]' ) self.assertEqual(vocab_keys[1] ,'[CLS]' ) self.assertEqual(vocab_keys[-1] ,'j' ) self.assertEqual(len(lowerCamelCase__ ) ,1_012 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_012 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) UpperCAmelCase__ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] ,) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] ,) @cached_property def __lowerCAmelCase ( self : Dict ): return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'Hello World!' UpperCAmelCase__ = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/xprophetnet-large-wiki100-cased' ,revision='1acad1643ddd54a44df6a1b797ada8373685d90e' ,)
632
0
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "encoder-decoder" snake_case__ = True def __init__( self : str ,**lowerCamelCase__ : Any ): super().__init__(**lowerCamelCase__ ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" UpperCAmelCase__ = kwargs.pop('encoder' ) UpperCAmelCase__ = encoder_config.pop('model_type' ) UpperCAmelCase__ = kwargs.pop('decoder' ) UpperCAmelCase__ = decoder_config.pop('model_type' ) from ..auto.configuration_auto import AutoConfig UpperCAmelCase__ = AutoConfig.for_model(lowerCamelCase__ ,**lowerCamelCase__ ) UpperCAmelCase__ = AutoConfig.for_model(lowerCamelCase__ ,**lowerCamelCase__ ) UpperCAmelCase__ = True @classmethod def __lowerCAmelCase ( cls : List[str] ,lowerCamelCase__ : PretrainedConfig ,lowerCamelCase__ : PretrainedConfig ,**lowerCamelCase__ : Dict ): logger.info('Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) UpperCAmelCase__ = True UpperCAmelCase__ = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ = self.encoder.to_dict() UpperCAmelCase__ = self.decoder.to_dict() UpperCAmelCase__ = self.__class__.model_type return output
720
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def a_ ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def a_ ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] lowerCAmelCase__ : str = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') lowerCAmelCase__ : str = sorted({word.strip().lower() for word in data.splitlines()}) lowerCAmelCase__ : Optional[Any] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": lowerCAmelCase__ : int = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
632
0
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class snake_case ( __UpperCAmelCase ): """simple docstring""" def __get__( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) UpperCAmelCase__ = '__cached_' + self.fget.__name__ UpperCAmelCase__ = getattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) if cached is None: UpperCAmelCase__ = self.fget(lowerCamelCase__ ) setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return cached def a_ ( lowerCamelCase ): UpperCAmelCase__ = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def a_ ( lowerCamelCase ): if is_torch_fx_proxy(lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return _is_numpy(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.Tensor ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.device ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_device(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch if isinstance(lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) else: return False return isinstance(lowerCamelCase , torch.dtype ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_dtype(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf return isinstance(lowerCamelCase , tf.Tensor ) def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tensorflow(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowerCamelCase , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowerCamelCase ) return type(lowerCamelCase ) == tf.Tensor def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowerCamelCase ) def a_ ( lowerCamelCase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowerCamelCase , jnp.ndarray ) def a_ ( lowerCamelCase ): return False if not is_flax_available() else _is_jax(lowerCamelCase ) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return [to_py_obj(lowerCamelCase ) for o in obj] elif is_tf_tensor(lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ).tolist() elif isinstance(lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return np.array(lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): return obj.numpy() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ) else: return obj class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = fields(self ) # Safety and consistency checks if not len(lowerCamelCase__ ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) UpperCAmelCase__ = getattr(self ,class_fields[0].name ) UpperCAmelCase__ = all(getattr(self ,field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase__ ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = first_field.items() UpperCAmelCase__ = True else: try: UpperCAmelCase__ = iter(lowerCamelCase__ ) UpperCAmelCase__ = True except TypeError: UpperCAmelCase__ = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase__ ): if ( not isinstance(lowerCamelCase__ ,(list, tuple) ) or not len(lowerCamelCase__ ) == 2 or not isinstance(element[0] ,lowerCamelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute UpperCAmelCase__ = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self ,element[0] ,element[1] ) if element[1] is not None: UpperCAmelCase__ = element[1] elif first_field is not None: UpperCAmelCase__ = first_field else: for field in class_fields: UpperCAmelCase__ = getattr(self ,field.name ) if v is not None: UpperCAmelCase__ = v def __delitem__( self : List[Any] ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : Dict ): raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : int ,**lowerCamelCase__ : Optional[Any] ): raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : List[Any] ,*lowerCamelCase__ : int ,**lowerCamelCase__ : List[Any] ): raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Optional[int] ): raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[Any] ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __setitem__( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): # Will raise a KeyException if needed super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): return tuple(self[k] for k in self.keys() ) class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : str ,lowerCamelCase__ : Optional[int] ): raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "longest" snake_case__ = "max_length" snake_case__ = "do_not_pad" class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "pt" snake_case__ = "tf" snake_case__ = "np" snake_case__ = "jax" class snake_case : """simple docstring""" def __init__( self : int ,lowerCamelCase__ : List[ContextManager] ): UpperCAmelCase__ = context_managers UpperCAmelCase__ = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase__ ) def __exit__( self : List[Any] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Dict ): self.stack.__exit__(*lowerCamelCase__ ,**lowerCamelCase__ ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( lowerCamelCase ): UpperCAmelCase__ = model_class.__name__ UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( lowerCamelCase , lowerCamelCase = "" , lowerCamelCase = "." ): def _flatten_dict(lowerCamelCase , lowerCamelCase="" , lowerCamelCase="." ): for k, v in d.items(): UpperCAmelCase__ = str(lowerCamelCase ) + delimiter + str(lowerCamelCase ) if parent_key else k if v and isinstance(lowerCamelCase , lowerCamelCase ): yield from flatten_dict(lowerCamelCase , lowerCamelCase , delimiter=lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) @contextmanager def a_ ( lowerCamelCase , lowerCamelCase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.transpose(lowerCamelCase , axes=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.T if axes is None else array.permute(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.transpose(lowerCamelCase , perm=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.transpose(lowerCamelCase , axes=lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.reshape(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.reshape(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.reshape(lowerCamelCase , lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.reshape(lowerCamelCase , lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.squeeze(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.expand_dims(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.unsqueeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.expand_dims(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.expand_dims(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.size(lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.numel() elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.size(lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): for key, value in auto_map.items(): if isinstance(lowerCamelCase , (tuple, list) ): UpperCAmelCase__ = [f'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: UpperCAmelCase__ = f'''{repo_id}--{value}''' return auto_map def a_ ( lowerCamelCase ): for base_class in inspect.getmro(lowerCamelCase ): UpperCAmelCase__ = base_class.__module__ UpperCAmelCase__ = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
721
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class snake_case ( ctypes.Structure ): """simple docstring""" snake_case__ = [("size", ctypes.c_int), ("visible", ctypes.c_byte)] def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def a_ ( ): try: hide_cursor() yield finally: show_cursor()
632
0
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class snake_case ( unittest.TestCase ): """simple docstring""" @parameterized.expand([(None,), ('foo.json',)] ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : int ): UpperCAmelCase__ = GenerationConfig( do_sample=lowerCamelCase__ ,temperature=0.7 ,length_penalty=1.0 ,bad_words_ids=[[1, 2, 3], [4, 5]] ,) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ ,config_name=lowerCamelCase__ ) UpperCAmelCase__ = GenerationConfig.from_pretrained(lowerCamelCase__ ,config_name=lowerCamelCase__ ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample ,lowerCamelCase__ ) self.assertEqual(loaded_config.temperature ,0.7 ) self.assertEqual(loaded_config.length_penalty ,1.0 ) self.assertEqual(loaded_config.bad_words_ids ,[[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k ,50 ) self.assertEqual(loaded_config.max_length ,20 ) self.assertEqual(loaded_config.max_time ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = AutoConfig.from_pretrained('gpt2' ) UpperCAmelCase__ = GenerationConfig.from_model_config(lowerCamelCase__ ) UpperCAmelCase__ = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(lowerCamelCase__ ,lowerCamelCase__ ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id ,default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id ,model_config.eos_token_id ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = GenerationConfig() UpperCAmelCase__ = { 'max_new_tokens': 1_024, 'foo': 'bar', } UpperCAmelCase__ = copy.deepcopy(lowerCamelCase__ ) UpperCAmelCase__ = generation_config.update(**lowerCamelCase__ ) # update_kwargs was not modified (no side effects) self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens ,1_024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(lowerCamelCase__ ,{'foo': 'bar'} ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = GenerationConfig() UpperCAmelCase__ = 'bar' with tempfile.TemporaryDirectory('test-generation-config' ) as tmp_dir: generation_config.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = GenerationConfig.from_pretrained(lowerCamelCase__ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo ,'bar' ) UpperCAmelCase__ = GenerationConfig.from_model_config(lowerCamelCase__ ) assert not hasattr(lowerCamelCase__ ,'foo' ) # no new kwargs should be initialized if from config def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = GenerationConfig() self.assertEqual(default_config.temperature ,1.0 ) self.assertEqual(default_config.do_sample ,lowerCamelCase__ ) self.assertEqual(default_config.num_beams ,1 ) UpperCAmelCase__ = GenerationConfig( do_sample=lowerCamelCase__ ,temperature=0.7 ,length_penalty=1.0 ,bad_words_ids=[[1, 2, 3], [4, 5]] ,) self.assertEqual(config.temperature ,0.7 ) self.assertEqual(config.do_sample ,lowerCamelCase__ ) self.assertEqual(config.num_beams ,1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = GenerationConfig.from_pretrained(lowerCamelCase__ ,temperature=1.0 ) self.assertEqual(loaded_config.temperature ,1.0 ) self.assertEqual(loaded_config.do_sample ,lowerCamelCase__ ) self.assertEqual(loaded_config.num_beams ,1 ) # default value @is_staging_test class snake_case ( unittest.TestCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : Optional[int] ): UpperCAmelCase__ = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ): try: delete_repo(token=cls._token ,repo_id='test-generation-config' ) except HTTPError: pass try: delete_repo(token=cls._token ,repo_id='valid_org/test-generation-config-org' ) except HTTPError: pass def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = GenerationConfig( do_sample=lowerCamelCase__ ,temperature=0.7 ,length_penalty=1.0 ,) config.push_to_hub('test-generation-config' ,use_auth_token=self._token ) UpperCAmelCase__ = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ ,getattr(lowerCamelCase__ ,lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token ,repo_id='test-generation-config' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ ,repo_id='test-generation-config' ,push_to_hub=lowerCamelCase__ ,use_auth_token=self._token ) UpperCAmelCase__ = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ ,getattr(lowerCamelCase__ ,lowerCamelCase__ ) ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = GenerationConfig( do_sample=lowerCamelCase__ ,temperature=0.7 ,length_penalty=1.0 ,) config.push_to_hub('valid_org/test-generation-config-org' ,use_auth_token=self._token ) UpperCAmelCase__ = GenerationConfig.from_pretrained('valid_org/test-generation-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ ,getattr(lowerCamelCase__ ,lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token ,repo_id='valid_org/test-generation-config-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ ,repo_id='valid_org/test-generation-config-org' ,push_to_hub=lowerCamelCase__ ,use_auth_token=self._token ) UpperCAmelCase__ = GenerationConfig.from_pretrained('valid_org/test-generation-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ ,getattr(lowerCamelCase__ ,lowerCamelCase__ ) )
700
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class snake_case ( __UpperCAmelCase ): """simple docstring""" def __get__( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) UpperCAmelCase__ = '__cached_' + self.fget.__name__ UpperCAmelCase__ = getattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) if cached is None: UpperCAmelCase__ = self.fget(lowerCamelCase__ ) setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return cached def a_ ( lowerCamelCase ): UpperCAmelCase__ = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def a_ ( lowerCamelCase ): if is_torch_fx_proxy(lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return _is_numpy(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.Tensor ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.device ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_device(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch if isinstance(lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) else: return False return isinstance(lowerCamelCase , torch.dtype ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_dtype(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf return isinstance(lowerCamelCase , tf.Tensor ) def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tensorflow(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowerCamelCase , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowerCamelCase ) return type(lowerCamelCase ) == tf.Tensor def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowerCamelCase ) def a_ ( lowerCamelCase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowerCamelCase , jnp.ndarray ) def a_ ( lowerCamelCase ): return False if not is_flax_available() else _is_jax(lowerCamelCase ) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return [to_py_obj(lowerCamelCase ) for o in obj] elif is_tf_tensor(lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ).tolist() elif isinstance(lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return np.array(lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): return obj.numpy() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ) else: return obj class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = fields(self ) # Safety and consistency checks if not len(lowerCamelCase__ ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) UpperCAmelCase__ = getattr(self ,class_fields[0].name ) UpperCAmelCase__ = all(getattr(self ,field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase__ ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = first_field.items() UpperCAmelCase__ = True else: try: UpperCAmelCase__ = iter(lowerCamelCase__ ) UpperCAmelCase__ = True except TypeError: UpperCAmelCase__ = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase__ ): if ( not isinstance(lowerCamelCase__ ,(list, tuple) ) or not len(lowerCamelCase__ ) == 2 or not isinstance(element[0] ,lowerCamelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute UpperCAmelCase__ = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self ,element[0] ,element[1] ) if element[1] is not None: UpperCAmelCase__ = element[1] elif first_field is not None: UpperCAmelCase__ = first_field else: for field in class_fields: UpperCAmelCase__ = getattr(self ,field.name ) if v is not None: UpperCAmelCase__ = v def __delitem__( self : List[Any] ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : Dict ): raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : int ,**lowerCamelCase__ : Optional[Any] ): raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : List[Any] ,*lowerCamelCase__ : int ,**lowerCamelCase__ : List[Any] ): raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Optional[int] ): raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[Any] ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __setitem__( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): # Will raise a KeyException if needed super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): return tuple(self[k] for k in self.keys() ) class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : str ,lowerCamelCase__ : Optional[int] ): raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "longest" snake_case__ = "max_length" snake_case__ = "do_not_pad" class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "pt" snake_case__ = "tf" snake_case__ = "np" snake_case__ = "jax" class snake_case : """simple docstring""" def __init__( self : int ,lowerCamelCase__ : List[ContextManager] ): UpperCAmelCase__ = context_managers UpperCAmelCase__ = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase__ ) def __exit__( self : List[Any] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Dict ): self.stack.__exit__(*lowerCamelCase__ ,**lowerCamelCase__ ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( lowerCamelCase ): UpperCAmelCase__ = model_class.__name__ UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( lowerCamelCase , lowerCamelCase = "" , lowerCamelCase = "." ): def _flatten_dict(lowerCamelCase , lowerCamelCase="" , lowerCamelCase="." ): for k, v in d.items(): UpperCAmelCase__ = str(lowerCamelCase ) + delimiter + str(lowerCamelCase ) if parent_key else k if v and isinstance(lowerCamelCase , lowerCamelCase ): yield from flatten_dict(lowerCamelCase , lowerCamelCase , delimiter=lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) @contextmanager def a_ ( lowerCamelCase , lowerCamelCase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.transpose(lowerCamelCase , axes=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.T if axes is None else array.permute(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.transpose(lowerCamelCase , perm=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.transpose(lowerCamelCase , axes=lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.reshape(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.reshape(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.reshape(lowerCamelCase , lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.reshape(lowerCamelCase , lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.squeeze(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.expand_dims(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.unsqueeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.expand_dims(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.expand_dims(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.size(lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.numel() elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.size(lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): for key, value in auto_map.items(): if isinstance(lowerCamelCase , (tuple, list) ): UpperCAmelCase__ = [f'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: UpperCAmelCase__ = f'''{repo_id}--{value}''' return auto_map def a_ ( lowerCamelCase ): for base_class in inspect.getmro(lowerCamelCase ): UpperCAmelCase__ = base_class.__module__ UpperCAmelCase__ = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
632
0
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
701
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ : Union[str, Any] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained(lowerCamelCase__ ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = generator.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'cyberpunk 2077' UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = 'A painting of a squirrel eating a burger ' UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.text_to_image( prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = pipe.image_variation(lowerCamelCase__ ,generator=lowerCamelCase__ ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
632
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : List[Any] = logging.get_logger(__name__) def a_ ( lowerCamelCase ): UpperCAmelCase__ = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): UpperCAmelCase__ = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): UpperCAmelCase__ = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase__ = key[key.find('patch_embed' ) + len('patch_embed' )] UpperCAmelCase__ = key.replace(f'''patch_embed{idx}''' , f'''patch_embeddings.{int(lowerCamelCase )-1}''' ) if "norm" in key: UpperCAmelCase__ = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase__ = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] UpperCAmelCase__ = key.replace(f'''layer_norm{idx}''' , f'''layer_norm.{int(lowerCamelCase )-1}''' ) if "layer_norm1" in key: UpperCAmelCase__ = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: UpperCAmelCase__ = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase__ = key[key.find('block' ) + len('block' )] UpperCAmelCase__ = key.replace(f'''block{idx}''' , f'''block.{int(lowerCamelCase )-1}''' ) if "attn.q" in key: UpperCAmelCase__ = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: UpperCAmelCase__ = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: UpperCAmelCase__ = key.replace('attn' , 'attention.self' ) if "fc1" in key: UpperCAmelCase__ = key.replace('fc1' , 'dense1' ) if "fc2" in key: UpperCAmelCase__ = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: UpperCAmelCase__ = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: UpperCAmelCase__ = key.replace('linear_fuse.conv' , 'linear_fuse' ) UpperCAmelCase__ = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase__ = key[key.find('linear_c' ) + len('linear_c' )] UpperCAmelCase__ = key.replace(f'''linear_c{idx}''' , f'''linear_c.{int(lowerCamelCase )-1}''' ) if "bot_conv" in key: UpperCAmelCase__ = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: UpperCAmelCase__ = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: UpperCAmelCase__ = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: UpperCAmelCase__ = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: UpperCAmelCase__ = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: UpperCAmelCase__ = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: UpperCAmelCase__ = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): UpperCAmelCase__ = key.replace('module.last_layer_depth' , 'head.head' ) UpperCAmelCase__ = value return new_state_dict def a_ ( lowerCamelCase , lowerCamelCase ): # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase__ = state_dict.pop(f'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase__ = state_dict.pop(f'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase__ = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase__ = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase__ = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase__ = kv_bias[config.hidden_sizes[i] :] def a_ ( ): UpperCAmelCase__ = 'http://images.cocodataset.org/val2017/000000039769.jpg' UpperCAmelCase__ = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ) return image @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase=False , lowerCamelCase=None ): UpperCAmelCase__ = GLPNConfig(hidden_sizes=[6_4, 1_2_8, 3_2_0, 5_1_2] , decoder_hidden_size=6_4 , depths=[3, 8, 2_7, 3] ) # load image processor (only resize + rescale) UpperCAmelCase__ = GLPNImageProcessor() # prepare image UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=lowerCamelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict UpperCAmelCase__ = torch.load(lowerCamelCase , map_location=torch.device('cpu' ) ) # rename keys UpperCAmelCase__ = rename_keys(lowerCamelCase ) # key and value matrices need special treatment read_in_k_v(lowerCamelCase , lowerCamelCase ) # create HuggingFace model and load state dict UpperCAmelCase__ = GLPNForDepthEstimation(lowerCamelCase ) model.load_state_dict(lowerCamelCase ) model.eval() # forward pass UpperCAmelCase__ = model(lowerCamelCase ) UpperCAmelCase__ = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase__ = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase__ = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(f'''Unknown model name: {model_name}''' ) UpperCAmelCase__ = torch.Size([1, 4_8_0, 6_4_0] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , lowerCamelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(lowerCamelCase , lowerCamelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=lowerCamelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(lowerCamelCase , lowerCamelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=lowerCamelCase , ) if __name__ == "__main__": lowerCAmelCase__ : int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) parser.add_argument( '--model_name', default='glpn-kitti', type=str, help='Name of the model in case you\'re pushing to the hub.', ) lowerCAmelCase__ : str = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
702
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
632
0
"""simple docstring""" from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer lowerCAmelCase__ : int = logging.get_logger(__name__) lowerCAmelCase__ : Optional[int] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase__ : List[Any] = { 'vocab_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json' }, 'merges_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt' }, } lowerCAmelCase__ : Optional[Any] = {'allegro/herbert-base-cased': 514} lowerCAmelCase__ : int = {} class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_INIT_CONFIGURATION snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = HerbertTokenizer def __init__( self : Union[str, Any] ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : Optional[int]="<s>" ,lowerCamelCase__ : Dict="<unk>" ,lowerCamelCase__ : Optional[Any]="<pad>" ,lowerCamelCase__ : List[Any]="<mask>" ,lowerCamelCase__ : List[str]="</s>" ,**lowerCamelCase__ : Tuple ,): super().__init__( lowerCamelCase__ ,lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,**lowerCamelCase__ ,) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): UpperCAmelCase__ = [self.cls_token_id] UpperCAmelCase__ = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ,lowerCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ ,token_ids_a=lowerCamelCase__ ,already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1] + ([0] * len(lowerCamelCase__ )) + [1] def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): UpperCAmelCase__ = self._tokenizer.model.save(lowerCamelCase__ ,name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
703
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
632
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__ : Union[str, Any] = { 'configuration_llama': ['LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LlamaConfig'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : int = ['LlamaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : str = ['LlamaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'LlamaForCausalLM', 'LlamaModel', 'LlamaPreTrainedModel', 'LlamaForSequenceClassification', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys lowerCAmelCase__ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
704
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
632
0
"""simple docstring""" from statistics import mean import numpy as np def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = 0 # Number of processes finished UpperCAmelCase__ = 0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. UpperCAmelCase__ = [0] * no_of_process # List to include calculation results UpperCAmelCase__ = [0] * no_of_process # Sort by arrival time. UpperCAmelCase__ = [burst_time[i] for i in np.argsort(lowerCamelCase )] UpperCAmelCase__ = [process_name[i] for i in np.argsort(lowerCamelCase )] arrival_time.sort() while no_of_process > finished_process_count: UpperCAmelCase__ = 0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: UpperCAmelCase__ = arrival_time[i] UpperCAmelCase__ = 0 # Index showing the location of the process being performed UpperCAmelCase__ = 0 # Saves the current response ratio. UpperCAmelCase__ = 0 for i in range(0 , lowerCamelCase ): if finished_process[i] == 0 and arrival_time[i] <= current_time: UpperCAmelCase__ = (burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: UpperCAmelCase__ = temp UpperCAmelCase__ = i # Calculate the turn around time UpperCAmelCase__ = current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. UpperCAmelCase__ = 1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = [0] * no_of_process for i in range(0 , lowerCamelCase ): UpperCAmelCase__ = turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": lowerCAmelCase__ : Union[str, Any] = 5 lowerCAmelCase__ : Dict = ['A', 'B', 'C', 'D', 'E'] lowerCAmelCase__ : Dict = [1, 2, 3, 4, 5] lowerCAmelCase__ : str = [1, 2, 3, 4, 5] lowerCAmelCase__ : int = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) lowerCAmelCase__ : Tuple = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print('Process name \tArrival time \tBurst time \tTurn around time \tWaiting time') for i in range(0, no_of_process): print( F"""{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t""" F"""{turn_around_time[i]}\t\t\t{waiting_time[i]}""" ) print(F"""average waiting time : {mean(waiting_time):.5f}""") print(F"""average turn around time : {mean(turn_around_time):.5f}""")
705
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
632
0
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCAmelCase__ : Optional[int] = '\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu")\n\n >>> repo = "openai/shap-e-img2img"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png"\n >>> image = load_image(image_url).convert("RGB")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], "corgi_3d.gif")\n ```\n' @dataclass class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = 42 class snake_case ( __UpperCAmelCase ): """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : PriorTransformer ,lowerCamelCase__ : CLIPVisionModel ,lowerCamelCase__ : CLIPImageProcessor ,lowerCamelCase__ : HeunDiscreteScheduler ,lowerCamelCase__ : ShapERenderer ,): super().__init__() self.register_modules( prior=lowerCamelCase__ ,image_encoder=lowerCamelCase__ ,image_processor=lowerCamelCase__ ,scheduler=lowerCamelCase__ ,renderer=lowerCamelCase__ ,) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[int] ): if latents is None: UpperCAmelCase__ = randn_tensor(lowerCamelCase__ ,generator=lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=lowerCamelCase__ ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) UpperCAmelCase__ = latents.to(lowerCamelCase__ ) UpperCAmelCase__ = latents * scheduler.init_noise_sigma return latents def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) UpperCAmelCase__ = torch.device(f'''cuda:{gpu_id}''' ) UpperCAmelCase__ = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase__ ,lowerCamelCase__ ) @property def __lowerCAmelCase ( self : str ): if self.device != torch.device('meta' ) or not hasattr(self.image_encoder ,'_hf_hook' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(lowerCamelCase__ ,'_hf_hook' ) and hasattr(module._hf_hook ,'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : int ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : List[Any] ,): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and isinstance(image[0] ,torch.Tensor ): UpperCAmelCase__ = torch.cat(lowerCamelCase__ ,axis=0 ) if image[0].ndim == 4 else torch.stack(lowerCamelCase__ ,axis=0 ) if not isinstance(lowerCamelCase__ ,torch.Tensor ): UpperCAmelCase__ = self.image_processor(lowerCamelCase__ ,return_tensors='pt' ).pixel_values[0].unsqueeze(0 ) UpperCAmelCase__ = image.to(dtype=self.image_encoder.dtype ,device=lowerCamelCase__ ) UpperCAmelCase__ = self.image_encoder(lowerCamelCase__ )['last_hidden_state'] UpperCAmelCase__ = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 UpperCAmelCase__ = image_embeds.repeat_interleave(lowerCamelCase__ ,dim=0 ) if do_classifier_free_guidance: UpperCAmelCase__ = torch.zeros_like(lowerCamelCase__ ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase__ = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(lowerCamelCase__ ) def __call__( self : str ,lowerCamelCase__ : Union[PIL.Image.Image, List[PIL.Image.Image]] ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : int = 25 ,lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,lowerCamelCase__ : Optional[torch.FloatTensor] = None ,lowerCamelCase__ : float = 4.0 ,lowerCamelCase__ : int = 64 ,lowerCamelCase__ : Optional[str] = "pil" ,lowerCamelCase__ : bool = True ,): if isinstance(lowerCamelCase__ ,PIL.Image.Image ): UpperCAmelCase__ = 1 elif isinstance(lowerCamelCase__ ,torch.Tensor ): UpperCAmelCase__ = image.shape[0] elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and isinstance(image[0] ,(torch.Tensor, PIL.Image.Image) ): UpperCAmelCase__ = len(lowerCamelCase__ ) else: raise ValueError( f'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(lowerCamelCase__ )}''' ) UpperCAmelCase__ = self._execution_device UpperCAmelCase__ = batch_size * num_images_per_prompt UpperCAmelCase__ = guidance_scale > 1.0 UpperCAmelCase__ = self._encode_image(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # prior self.scheduler.set_timesteps(lowerCamelCase__ ,device=lowerCamelCase__ ) UpperCAmelCase__ = self.scheduler.timesteps UpperCAmelCase__ = self.prior.config.num_embeddings UpperCAmelCase__ = self.prior.config.embedding_dim UpperCAmelCase__ = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) ,image_embeds.dtype ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,self.scheduler ,) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim UpperCAmelCase__ = latents.reshape(latents.shape[0] ,lowerCamelCase__ ,lowerCamelCase__ ) for i, t in enumerate(self.progress_bar(lowerCamelCase__ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase__ = self.scheduler.scale_model_input(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self.prior( lowerCamelCase__ ,timestep=lowerCamelCase__ ,proj_embedding=lowerCamelCase__ ,).predicted_image_embedding # remove the variance UpperCAmelCase__ , UpperCAmelCase__ = noise_pred.split( scaled_model_input.shape[2] ,dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: UpperCAmelCase__ , UpperCAmelCase__ = noise_pred.chunk(2 ) UpperCAmelCase__ = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) UpperCAmelCase__ = self.scheduler.step( lowerCamelCase__ ,timestep=lowerCamelCase__ ,sample=lowerCamelCase__ ,).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=lowerCamelCase__ ) UpperCAmelCase__ = [] for i, latent in enumerate(lowerCamelCase__ ): print() UpperCAmelCase__ = self.renderer.decode( latent[None, :] ,lowerCamelCase__ ,size=lowerCamelCase__ ,ray_batch_size=4_096 ,n_coarse_samples=64 ,n_fine_samples=128 ,) images.append(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) if output_type not in ["np", "pil"]: raise ValueError(f'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) UpperCAmelCase__ = images.cpu().numpy() if output_type == "pil": UpperCAmelCase__ = [self.numpy_to_pil(lowerCamelCase__ ) for image in images] # Offload last model to CPU if hasattr(self ,'final_offload_hook' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=lowerCamelCase__ )
706
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ : int = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Union[str, Any] = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
632
0
"""simple docstring""" import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def a_ ( lowerCamelCase=None ): if subparsers is not None: UpperCAmelCase__ = subparsers.add_parser('env' ) else: UpperCAmelCase__ = argparse.ArgumentParser('Accelerate env command' ) parser.add_argument( '--config_file' , default=lowerCamelCase , help='The config file to use for the default values in the launching script.' ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def a_ ( lowerCamelCase ): UpperCAmelCase__ = torch.__version__ UpperCAmelCase__ = torch.cuda.is_available() UpperCAmelCase__ = is_xpu_available() UpperCAmelCase__ = is_npu_available() UpperCAmelCase__ = 'Not found' # Get the default from the config file. if args.config_file is not None or os.path.isfile(lowerCamelCase ): UpperCAmelCase__ = load_config_from_file(args.config_file ).to_dict() UpperCAmelCase__ = { '`Accelerate` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'Numpy version': np.__version__, 'PyTorch version (GPU?)': f'''{pt_version} ({pt_cuda_available})''', 'PyTorch XPU available': str(lowerCamelCase ), 'PyTorch NPU available': str(lowerCamelCase ), 'System RAM': f'''{psutil.virtual_memory().total / 1_0_2_4 ** 3:.2f} GB''', } if pt_cuda_available: UpperCAmelCase__ = torch.cuda.get_device_name() print('\nCopy-and-paste the text below in your GitHub issue\n' ) print('\n'.join([f'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('- `Accelerate` default config:' if args.config_file is None else '- `Accelerate` config passed:' ) UpperCAmelCase__ = ( '\n'.join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(lowerCamelCase , lowerCamelCase ) else f'''\t{accelerate_config}''' ) print(lowerCamelCase ) UpperCAmelCase__ = accelerate_config return info def a_ ( ): UpperCAmelCase__ = env_command_parser() UpperCAmelCase__ = parser.parse_args() env_command(lowerCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
707
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase__ : Optional[int] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( image=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
632
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_50, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_00, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_00, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ] ) class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): if self.framework == "pytorch": subprocess.run( f'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() ,encoding='utf-8' ,check=lowerCamelCase__ ,) assert hasattr(self ,'env' ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Dict ): UpperCAmelCase__ = f'''{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}''' # distributed data settings UpperCAmelCase__ = {'smdistributed': {'dataparallel': {'enabled': True}}} if self.script != 'run_ddp.py' else None # creates estimator return HuggingFace( entry_point=self.script ,source_dir=self.env.test_path ,role=self.env.role ,image_uri=self.env.image_uri ,base_job_name=lowerCamelCase__ ,instance_count=lowerCamelCase__ ,instance_type=self.instance_type ,debugger_hook_config=lowerCamelCase__ ,hyperparameters={**self.env.distributed_hyperparameters, 'model_name_or_path': self.model_name_or_path} ,metric_definitions=self.env.metric_definitions ,distribution=lowerCamelCase__ ,py_version='py36' ,) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Dict ): TrainingJobAnalytics(lowerCamelCase__ ).export_csv(f'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(2,)] ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : Optional[int] ): # create estimator UpperCAmelCase__ = self.create_estimator(lowerCamelCase__ ) # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == 'eval_accuracy']['value'] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == 'eval_loss']['value'] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get('TrainingTimeInSeconds' ,999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['eval_accuracy'] for t in eval_accuracy ) assert all(t <= self.results['eval_loss'] for t in eval_loss ) # dump tests result into json file to share in PR with open(f'''{estimator.latest_training_job.name}.json''' ,'w' ) as outfile: json.dump({'train_time': train_runtime, 'eval_accuracy': eval_accuracy, 'eval_loss': eval_loss} ,lowerCamelCase__ )
708
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def a_ ( lowerCamelCase = 2_0_0_0_0_0_0 ): UpperCAmelCase__ = [0] UpperCAmelCase__ = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target UpperCAmelCase__ = 0 # the area corresponding to the grid that gives the product closest to target UpperCAmelCase__ = 0 # an estimate of b, using the quadratic formula UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the triangle number corresponding to b_floor UpperCAmelCase__ = 42 # the triangle number corresponding to b_ceil UpperCAmelCase__ = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): UpperCAmelCase__ = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 UpperCAmelCase__ = floor(lowerCamelCase ) UpperCAmelCase__ = ceil(lowerCamelCase ) UpperCAmelCase__ = triangle_numbers[b_floor] UpperCAmelCase__ = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_first_guess * triangle_a UpperCAmelCase__ = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_second_guess * triangle_a UpperCAmelCase__ = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
632
0
from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ["input_features", "attention_mask"] def __init__( self : Optional[int] ,lowerCamelCase__ : int=80 ,lowerCamelCase__ : Any=16_000 ,lowerCamelCase__ : int=80 ,lowerCamelCase__ : str=0.0 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=True ,lowerCamelCase__ : Optional[Any]=True ,**lowerCamelCase__ : str ,): super().__init__(feature_size=lowerCamelCase__ ,sampling_rate=lowerCamelCase__ ,padding_value=lowerCamelCase__ ,**lowerCamelCase__ ) UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = do_ceptral_normalize UpperCAmelCase__ = normalize_means UpperCAmelCase__ = normalize_vars UpperCAmelCase__ = True def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : np.ndarray ,): UpperCAmelCase__ = waveform * (2**15) # Kaldi compliance: 16-bit signed integers UpperCAmelCase__ = torch.from_numpy(lowerCamelCase__ ).unsqueeze(0 ) UpperCAmelCase__ = ta_kaldi.fbank(lowerCamelCase__ ,num_mel_bins=self.num_mel_bins ,sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : float = 0.0 ,): # make sure we normalize float32 arrays if normalize_means: UpperCAmelCase__ = x[:input_length].mean(axis=0 ) UpperCAmelCase__ = np.subtract(lowerCamelCase__ ,lowerCamelCase__ ) if normalize_vars: UpperCAmelCase__ = x[:input_length].std(axis=0 ) UpperCAmelCase__ = np.divide(lowerCamelCase__ ,lowerCamelCase__ ) if input_length < x.shape[0]: UpperCAmelCase__ = padding_value # make sure array is in float32 UpperCAmelCase__ = x.astype(np.floataa ) return x def __lowerCAmelCase ( self : int ,lowerCamelCase__ : List[np.ndarray] ,lowerCamelCase__ : Optional[np.ndarray] = None ): UpperCAmelCase__ = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(lowerCamelCase__ ,lowerCamelCase__ ,self.normalize_means ,self.normalize_vars ,self.padding_value ) for x, n in zip(lowerCamelCase__ ,lowerCamelCase__ ) ] def __call__( self : Dict ,lowerCamelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,**lowerCamelCase__ : Union[str, Any] ,): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) UpperCAmelCase__ = isinstance(lowerCamelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(lowerCamelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCamelCase__ ,np.ndarray ): UpperCAmelCase__ = np.asarray(lowerCamelCase__ ,dtype=np.floataa ) elif isinstance(lowerCamelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [raw_speech] # extract fbank features UpperCAmelCase__ = [self._extract_fbank_features(lowerCamelCase__ ) for waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ = BatchFeature({'input_features': features} ) UpperCAmelCase__ = self.pad( lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,truncation=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,**lowerCamelCase__ ,) # make sure list is in array format UpperCAmelCase__ = padded_inputs.get('input_features' ) if isinstance(input_features[0] ,lowerCamelCase__ ): UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ = padded_inputs.get('attention_mask' ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ,dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: UpperCAmelCase__ = ( np.array(lowerCamelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCamelCase__ ,max_length=lowerCamelCase__ ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.normalize( padded_inputs['input_features'] ,attention_mask=lowerCamelCase__ ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(lowerCamelCase__ ) return padded_inputs
709
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : Optional[Any] = ['model.decoder.embed_positions.weights'] def a_ ( lowerCamelCase ): if "emb" in name: UpperCAmelCase__ = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: UpperCAmelCase__ = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: UpperCAmelCase__ = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: UpperCAmelCase__ = name.replace('linear1' , 'fc1' ) if "linear2" in name: UpperCAmelCase__ = name.replace('linear2' , 'fc2' ) if "norm1" in name: UpperCAmelCase__ = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: UpperCAmelCase__ = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: UpperCAmelCase__ = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: UpperCAmelCase__ = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: UpperCAmelCase__ = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase__ = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = list(state_dict.keys() ) UpperCAmelCase__ = {} for key in keys: UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) UpperCAmelCase__ = rename_keys(lowerCamelCase ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase__ = val[:hidden_size, :] UpperCAmelCase__ = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase__ = val else: UpperCAmelCase__ = val return state_dict, enc_dec_proj_state_dict def a_ ( lowerCamelCase ): if checkpoint == "small": # default config values UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif checkpoint == "medium": UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 2_4 elif checkpoint == "large": UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 3_2 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) UpperCAmelCase__ = MusicgenDecoderConfig( hidden_size=lowerCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=lowerCamelCase , num_attention_heads=lowerCamelCase , ) return config @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="cpu" ): UpperCAmelCase__ = MusicGen.get_pretrained(lowerCamelCase , device=lowerCamelCase ) UpperCAmelCase__ = decoder_config_from_checkpoint(lowerCamelCase ) UpperCAmelCase__ = fairseq_model.lm.state_dict() UpperCAmelCase__ , UpperCAmelCase__ = rename_state_dict( lowerCamelCase , hidden_size=decoder_config.hidden_size ) UpperCAmelCase__ = TaEncoderModel.from_pretrained('t5-base' ) UpperCAmelCase__ = EncodecModel.from_pretrained('facebook/encodec_32khz' ) UpperCAmelCase__ = MusicgenForCausalLM(lowerCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase__ , UpperCAmelCase__ = decoder.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model UpperCAmelCase__ = MusicgenForConditionalGeneration(text_encoder=lowerCamelCase , audio_encoder=lowerCamelCase , decoder=lowerCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(lowerCamelCase ) # check we can do a forward pass UpperCAmelCase__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase__ = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits if logits.shape != (8, 1, 2_0_4_8): raise ValueError('Incorrect shape for logits' ) # now construct the processor UpperCAmelCase__ = AutoTokenizer.from_pretrained('t5-base' ) UpperCAmelCase__ = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) UpperCAmelCase__ = MusicgenProcessor(feature_extractor=lowerCamelCase , tokenizer=lowerCamelCase ) # set the appropriate bos/pad token ids UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 2_0_4_8 # set other default generation config params UpperCAmelCase__ = int(3_0 * audio_encoder.config.frame_rate ) UpperCAmelCase__ = True UpperCAmelCase__ = 3.0 if pytorch_dump_folder is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(lowerCamelCase ) processor.save_pretrained(lowerCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(lowerCamelCase ) processor.push_to_hub(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
632
0
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments lowerCAmelCase__ : Dict = logging.getLogger(__name__) @dataclass class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = field( default=0.0 , metadata={"help": "The label smoothing epsilon to apply (if not zero)."} ) snake_case__ = field(default=__UpperCAmelCase , metadata={"help": "Whether to SortishSamler or not."} ) snake_case__ = field( default=__UpperCAmelCase , metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) snake_case__ = field(default=__UpperCAmelCase , metadata={"help": "whether to use adafactor"} ) snake_case__ = field( default=__UpperCAmelCase , metadata={"help": "Encoder layer dropout probability. Goes into model.config."} ) snake_case__ = field( default=__UpperCAmelCase , metadata={"help": "Decoder layer dropout probability. Goes into model.config."} ) snake_case__ = field(default=__UpperCAmelCase , metadata={"help": "Dropout probability. Goes into model.config."} ) snake_case__ = field( default=__UpperCAmelCase , metadata={"help": "Attention dropout probability. Goes into model.config."} ) snake_case__ = field( default="linear" , metadata={"help": F'Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'} , )
710
"""simple docstring""" lowerCAmelCase__ : Tuple = range(2, 20 + 1) lowerCAmelCase__ : Optional[Any] = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase__ : dict[int, dict[int, list[list[int]]]] = {} def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(a_i[j] for j in range(lowerCamelCase , len(lowerCamelCase ) ) ) UpperCAmelCase__ = sum(a_i[j] * base[j] for j in range(min(len(lowerCamelCase ) , lowerCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ = 0, 0 UpperCAmelCase__ = n - i UpperCAmelCase__ = memo.get(lowerCamelCase ) if sub_memo is not None: UpperCAmelCase__ = sub_memo.get(lowerCamelCase ) if jumps is not None and len(lowerCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ = -1 for _k in range(len(lowerCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ = diff + c for j in range(min(lowerCamelCase , len(lowerCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) if new_c > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: UpperCAmelCase__ = [] else: UpperCAmelCase__ = {c: []} UpperCAmelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , k - 1 , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ = compute(lowerCamelCase , lowerCamelCase , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ = 0 while j < len(lowerCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowerCamelCase , (diff, dn, k) ) return (diff, dn) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if i >= n: return 0, i if k > len(lowerCamelCase ): a_i.extend([0 for _ in range(k - len(lowerCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 0, 0, 0 for j in range(len(lowerCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ = ds_c + ds_b diff += addend UpperCAmelCase__ = 0 for j in range(lowerCamelCase ): UpperCAmelCase__ = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return diff, i - start_i def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): for j in range(lowerCamelCase , len(lowerCamelCase ) ): UpperCAmelCase__ = digits[j] + addend if s >= 1_0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) UpperCAmelCase__ = addend // 1_0 + quotient else: UpperCAmelCase__ = s UpperCAmelCase__ = addend // 1_0 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) digits.append(lowerCamelCase ) def a_ ( lowerCamelCase = 1_0**1_5 ): UpperCAmelCase__ = [1] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , 2_0 , i + dn , lowerCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ = 0 for j in range(len(lowerCamelCase ) ): a_n += digits[j] * 1_0**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowerCAmelCase__ : Tuple = logging.getLogger() def a_ ( lowerCamelCase ): UpperCAmelCase__ = {} UpperCAmelCase__ = os.path.join(lowerCamelCase , 'all_results.json' ) if os.path.exists(lowerCamelCase ): with open(lowerCamelCase , 'r' ) as f: UpperCAmelCase__ = json.load(lowerCamelCase ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowerCAmelCase__ : Any = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): import xla_spawn UpperCAmelCase__ = self.get_auto_remove_tmp_dir() UpperCAmelCase__ = f''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowerCamelCase__ ,'argv' ,lowerCamelCase__ ): UpperCAmelCase__ = time() xla_spawn.main() UpperCAmelCase__ = time() UpperCAmelCase__ = get_results(lowerCamelCase__ ) self.assertGreaterEqual(result['eval_accuracy'] ,0.7_5 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start ,500 ) def __lowerCAmelCase ( self : str ): import xla_spawn UpperCAmelCase__ = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowerCamelCase__ ,'argv' ,lowerCamelCase__ ): xla_spawn.main()
711
"""simple docstring""" import random class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : str ): UpperCAmelCase__ = [ord(lowerCamelCase__ ) for i in text] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i in plain: UpperCAmelCase__ = random.randint(1 ,300 ) UpperCAmelCase__ = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): UpperCAmelCase__ = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ : Dict = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
632
0
"""simple docstring""" def a_ ( lowerCamelCase = 5_0 ): UpperCAmelCase__ = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F"""{solution() = }""")
712
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
632
0
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): # Construct model if openai_config_file == "": UpperCAmelCase__ = OpenAIGPTConfig() else: UpperCAmelCase__ = OpenAIGPTConfig.from_json_file(lowerCamelCase ) UpperCAmelCase__ = OpenAIGPTModel(lowerCamelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Save pytorch-model UpperCAmelCase__ = pytorch_dump_folder_path + '/' + WEIGHTS_NAME UpperCAmelCase__ = pytorch_dump_folder_path + '/' + CONFIG_NAME print(f'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict() , lowerCamelCase ) print(f'''Save configuration file to {pytorch_config_dump_path}''' ) with open(lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCAmelCase__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--openai_checkpoint_folder_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--openai_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) lowerCAmelCase__ : str = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
713
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Dict ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_torch def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,torch.tensor(lowerCamelCase__ ) ,torch.tensor(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) @require_vision @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_tf def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,tf.convert_to_tensor(lowerCamelCase__ ) ,tf.convert_to_tensor(lowerCamelCase__ ) ,return_tensors='tf' ,) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowerCamelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = np.random.randint(0 ,2 ,size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase__ = [tf.convert_to_tensor(lowerCamelCase__ )] UpperCAmelCase__ = [torch.tensor(lowerCamelCase__ )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='pt' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) )
632
0
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch lowerCAmelCase__ : Any = random.Random() def a_ ( lowerCamelCase , lowerCamelCase=1.0 , lowerCamelCase=None , lowerCamelCase=None ): if rng is None: UpperCAmelCase__ = global_rng UpperCAmelCase__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Optional[int]=400 ,lowerCamelCase__ : List[str]=2_000 ,lowerCamelCase__ : Any=1 ,lowerCamelCase__ : Any=0.0 ,lowerCamelCase__ : str=16_000 ,lowerCamelCase__ : str=True ,lowerCamelCase__ : Union[str, Any]=80 ,lowerCamelCase__ : List[Any]=16 ,lowerCamelCase__ : List[str]=64 ,lowerCamelCase__ : Union[str, Any]="hann_window" ,lowerCamelCase__ : Optional[Any]=80 ,lowerCamelCase__ : Dict=7_600 ,lowerCamelCase__ : List[Any]=1e-10 ,lowerCamelCase__ : Union[str, Any]=True ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = min_seq_length UpperCAmelCase__ = max_seq_length UpperCAmelCase__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCAmelCase__ = feature_size UpperCAmelCase__ = padding_value UpperCAmelCase__ = sampling_rate UpperCAmelCase__ = do_normalize UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = return_attention_mask def __lowerCAmelCase ( self : Tuple ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : Tuple=False ): def _flatten(lowerCamelCase__ : List[Any] ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: UpperCAmelCase__ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size UpperCAmelCase__ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Union[str, Any]=False ,lowerCamelCase__ : Optional[Any]=False ): if equal_length: UpperCAmelCase__ = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size UpperCAmelCase__ = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = SpeechTaFeatureExtractor def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = SpeechTaFeatureExtractionTester(self ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Optional[Any] ): self.assertTrue(np.all(np.mean(lowerCamelCase__ ,axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ,axis=0 ) - 1 ) < 1e-3 ) ) def __lowerCAmelCase ( self : Any ): # Tests that all call wrap to encode_plus and batch_encode_plus UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test not batched input UpperCAmelCase__ = feat_extract(speech_inputs[0] ,return_tensors='np' ).input_values UpperCAmelCase__ = feat_extract(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) # Test batched UpperCAmelCase__ = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values UpperCAmelCase__ = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = ['longest', 'max_length', 'do_not_pad'] UpperCAmelCase__ = [None, 1_600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = feat_extract(lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self.assertTrue(input_values[0][1_000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = range(800 ,1_400 ,200 ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in lengths] UpperCAmelCase__ = ['longest', 'max_length', 'do_not_pad'] UpperCAmelCase__ = [None, 1_600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = feat_extract(lowerCamelCase__ ,max_length=lowerCamelCase__ ,padding=lowerCamelCase__ ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1_000 ,padding='max_length' ,return_tensors='np' ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1_000 ,padding='longest' ,return_tensors='np' ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1_000) ) UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=2_000 ,padding='longest' ,return_tensors='np' ) UpperCAmelCase__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1_200) ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCAmelCase__ = np.random.rand(100 ).astype(np.floataa ) UpperCAmelCase__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCAmelCase__ = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) UpperCAmelCase__ = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __lowerCAmelCase ( self : str ): # Tests that all call wrap to encode_plus and batch_encode_plus UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCAmelCase__ = [floats_list((1, x) )[0] for x in range(800 ,1_400 ,200 )] UpperCAmelCase__ = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size UpperCAmelCase__ = feature_extractor(audio_target=lowerCamelCase__ ,padding=lowerCamelCase__ ,return_tensors='np' ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input UpperCAmelCase__ = feature_extractor(speech_inputs[0] ,return_tensors='np' ).input_values UpperCAmelCase__ = feature_extractor(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) # Test batched UpperCAmelCase__ = feature_extractor(lowerCamelCase__ ,return_tensors='np' ).input_values UpperCAmelCase__ = feature_extractor(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCAmelCase__ = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCAmelCase__ = np.asarray(lowerCamelCase__ ) UpperCAmelCase__ = feature_extractor(lowerCamelCase__ ,return_tensors='np' ).input_values UpperCAmelCase__ = feature_extractor(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target() UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) UpperCAmelCase__ = feat_extract.model_input_names[0] UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) for x, y in zip(lowerCamelCase__ ,processed_features[input_name] ) ) ) UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCamelCase__ ) UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) UpperCAmelCase__ = processed_features[input_name] if len(batch_features_input.shape ) < 3: UpperCAmelCase__ = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCamelCase__ ) UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) UpperCAmelCase__ = feat_extract.model_input_names[0] UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) UpperCAmelCase__ = processed_features[input_name] if len(batch_features_input.shape ) < 3: UpperCAmelCase__ = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target() UpperCAmelCase__ = feat_extract.model_input_names[0] UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ) UpperCAmelCase__ = feat_extract.num_mel_bins # hack! UpperCAmelCase__ = feat_extract.pad(lowerCamelCase__ ,padding='longest' ,return_tensors='np' )[input_name] UpperCAmelCase__ = feat_extract.pad(lowerCamelCase__ ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.feat_extract_dict UpperCAmelCase__ = True UpperCAmelCase__ = self.feature_extraction_class(**lowerCamelCase__ ) UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target() UpperCAmelCase__ = [len(lowerCamelCase__ ) for x in speech_inputs] UpperCAmelCase__ = feat_extract.model_input_names[0] UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ) UpperCAmelCase__ = feat_extract.num_mel_bins # hack! UpperCAmelCase__ = feat_extract.pad(lowerCamelCase__ ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,lowerCamelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = self.feat_extract_dict UpperCAmelCase__ = True UpperCAmelCase__ = self.feature_extraction_class(**lowerCamelCase__ ) UpperCAmelCase__ = self.feat_extract_tester.prepare_inputs_for_target() UpperCAmelCase__ = [len(lowerCamelCase__ ) for x in speech_inputs] UpperCAmelCase__ = feat_extract.model_input_names[0] UpperCAmelCase__ = BatchFeature({input_name: speech_inputs} ) UpperCAmelCase__ = min(lowerCamelCase__ ) UpperCAmelCase__ = feat_extract.num_mel_bins # hack! UpperCAmelCase__ = feat_extract.pad( lowerCamelCase__ ,padding='max_length' ,max_length=lowerCamelCase__ ,truncation=lowerCamelCase__ ,return_tensors='np' ) self.assertIn('attention_mask' ,lowerCamelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Optional[int] ): from datasets import load_dataset UpperCAmelCase__ = load_dataset('hf-internal-testing/librispeech_asr_dummy' ,'clean' ,split='validation' ) # automatic decoding with librispeech UpperCAmelCase__ = ds.sort('id' ).select(range(lowerCamelCase__ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def __lowerCAmelCase ( self : Any ): # fmt: off UpperCAmelCase__ = torch.tensor( [2.3804e-03, 2.0752e-03, 1.9836e-03, 2.1057e-03, 1.6174e-03, 3.0518e-04, 9.1553e-05, 3.3569e-04, 9.7656e-04, 1.8311e-03, 2.0142e-03, 2.1057e-03, 1.7395e-03, 4.5776e-04, -3.9673e-04, 4.5776e-04, 1.0071e-03, 9.1553e-05, 4.8828e-04, 1.1597e-03, 7.3242e-04, 9.4604e-04, 1.8005e-03, 1.8311e-03, 8.8501e-04, 4.2725e-04, 4.8828e-04, 7.3242e-04, 1.0986e-03, 2.1057e-03] ) # fmt: on UpperCAmelCase__ = self._load_datasamples(1 ) UpperCAmelCase__ = SpeechTaFeatureExtractor() UpperCAmelCase__ = feature_extractor(lowerCamelCase__ ,return_tensors='pt' ).input_values self.assertEquals(input_values.shape ,(1, 93_680) ) self.assertTrue(torch.allclose(input_values[0, :30] ,lowerCamelCase__ ,atol=1e-6 ) ) def __lowerCAmelCase ( self : Optional[Any] ): # fmt: off UpperCAmelCase__ = torch.tensor( [-2.6_8_7_0, -3.0_1_0_4, -3.1_3_5_6, -3.5_3_5_2, -3.0_0_4_4, -3.0_3_5_3, -3.4_7_1_9, -3.6_7_7_7, -3.1_5_2_0, -2.9_4_3_5, -2.6_5_5_3, -2.8_7_9_5, -2.9_9_4_4, -2.5_9_2_1, -3.0_2_7_9, -3.0_3_8_6, -3.0_8_6_4, -3.1_2_9_1, -3.2_3_5_3, -2.7_4_4_4, -2.6_8_3_1, -2.7_2_8_7, -3.1_7_6_1, -3.1_5_7_1, -3.2_7_2_6, -3.0_5_8_2, -3.1_0_0_7, -3.4_5_3_3, -3.4_6_9_5, -3.0_9_9_8] ) # fmt: on UpperCAmelCase__ = self._load_datasamples(1 ) UpperCAmelCase__ = SpeechTaFeatureExtractor() UpperCAmelCase__ = feature_extractor(audio_target=lowerCamelCase__ ,return_tensors='pt' ).input_values self.assertEquals(input_values.shape ,(1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] ,lowerCamelCase__ ,atol=1e-4 ) )
714
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) lowerCAmelCase__ : str = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "ctrl" snake_case__ = ["past_key_values"] snake_case__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any ,lowerCamelCase__ : str=246_534 ,lowerCamelCase__ : List[str]=256 ,lowerCamelCase__ : Optional[int]=1_280 ,lowerCamelCase__ : Any=8_192 ,lowerCamelCase__ : int=48 ,lowerCamelCase__ : Optional[Any]=16 ,lowerCamelCase__ : Union[str, Any]=0.1 ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : List[str]=1e-6 ,lowerCamelCase__ : List[str]=0.0_2 ,lowerCamelCase__ : Tuple=True ,**lowerCamelCase__ : Optional[Any] ,): UpperCAmelCase__ = vocab_size UpperCAmelCase__ = n_positions UpperCAmelCase__ = n_embd UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = dff UpperCAmelCase__ = resid_pdrop UpperCAmelCase__ = embd_pdrop UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache super().__init__(**lowerCamelCase__ )
632
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class snake_case ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) UpperCAmelCase__ = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 25_543, 110, 83, 6]] ,dtype=tf.intaa ,) # J'aime le camembert !" UpperCAmelCase__ = model(lowerCamelCase__ )['last_hidden_state'] UpperCAmelCase__ = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape ,lowerCamelCase__ ) # compare the actual values for a slice. UpperCAmelCase__ = tf.convert_to_tensor( [[[-0.0_2_5_4, 0.0_2_3_5, 0.1_0_2_7], [0.0_6_0_6, -0.1_8_1_1, -0.0_4_1_8], [-0.1_5_6_1, -0.1_1_2_7, 0.2_6_8_7]]] ,dtype=tf.floataa ,) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1e-4 ) )
715
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCAmelCase__ : Dict = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def a_ ( ): UpperCAmelCase__ = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCAmelCase__ = get_sagemaker_input() else: UpperCAmelCase__ = get_cluster_input() return config def a_ ( lowerCamelCase=None ): if subparsers is not None: UpperCAmelCase__ = subparsers.add_parser('config' , description=lowerCamelCase ) else: UpperCAmelCase__ = argparse.ArgumentParser('Accelerate config command' , description=lowerCamelCase ) parser.add_argument( '--config_file' , default=lowerCamelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def a_ ( lowerCamelCase ): UpperCAmelCase__ = get_user_input() if args.config_file is not None: UpperCAmelCase__ = args.config_file else: if not os.path.isdir(lowerCamelCase ): os.makedirs(lowerCamelCase ) UpperCAmelCase__ = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(lowerCamelCase ) else: config.to_yaml_file(lowerCamelCase ) print(f'''accelerate configuration saved at {config_file}''' ) def a_ ( ): UpperCAmelCase__ = config_command_parser() UpperCAmelCase__ = parser.parse_args() config_command(lowerCamelCase ) if __name__ == "__main__": main()
632
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : List[str]=7 ,lowerCamelCase__ : List[Any]=3 ,lowerCamelCase__ : List[Any]=18 ,lowerCamelCase__ : Optional[int]=30 ,lowerCamelCase__ : Tuple=400 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Dict=None ,lowerCamelCase__ : Tuple=True ,): UpperCAmelCase__ = size if size is not None else {'height': 18, 'width': 18} UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = min_resolution UpperCAmelCase__ = max_resolution UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = apply_ocr def __lowerCAmelCase ( self : Dict ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = LayoutLMvaImageProcessor if is_pytesseract_available() else None def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = LayoutLMvaImageProcessingTester(self ) @property def __lowerCAmelCase ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ ,'do_resize' ) ) self.assertTrue(hasattr(lowerCamelCase__ ,'size' ) ) self.assertTrue(hasattr(lowerCamelCase__ ,'apply_ocr' ) ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'height': 18, 'width': 18} ) UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{'height': 42, 'width': 42} ) def __lowerCAmelCase ( self : int ): pass def __lowerCAmelCase ( self : List[str] ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,Image.Image ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ) self.assertEqual( encoding.pixel_values.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) self.assertIsInstance(encoding.words ,lowerCamelCase__ ) self.assertIsInstance(encoding.boxes ,lowerCamelCase__ ) # Test batched UpperCAmelCase__ = image_processing(lowerCamelCase__ ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) def __lowerCAmelCase ( self : Any ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ,numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,np.ndarray ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) # Test batched UpperCAmelCase__ = image_processing(lowerCamelCase__ ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) def __lowerCAmelCase ( self : Optional[Any] ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ,torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,torch.Tensor ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) # Test batched UpperCAmelCase__ = image_processing(lowerCamelCase__ ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) def __lowerCAmelCase ( self : Dict ): # with apply_OCR = True UpperCAmelCase__ = LayoutLMvaImageProcessor() from datasets import load_dataset UpperCAmelCase__ = load_dataset('hf-internal-testing/fixtures_docvqa' ,split='test' ) UpperCAmelCase__ = Image.open(ds[0]['file'] ).convert('RGB' ) UpperCAmelCase__ = image_processing(lowerCamelCase__ ,return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) ,len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 UpperCAmelCase__ = [['11:14', 'to', '11:39', 'a.m', '11:39', 'to', '11:44', 'a.m.', '11:44', 'a.m.', 'to', '12:25', 'p.m.', '12:25', 'to', '12:58', 'p.m.', '12:58', 'to', '4:00', 'p.m.', '2:00', 'to', '5:00', 'p.m.', 'Coffee', 'Break', 'Coffee', 'will', 'be', 'served', 'for', 'men', 'and', 'women', 'in', 'the', 'lobby', 'adjacent', 'to', 'exhibit', 'area.', 'Please', 'move', 'into', 'exhibit', 'area.', '(Exhibits', 'Open)', 'TRRF', 'GENERAL', 'SESSION', '(PART', '|)', 'Presiding:', 'Lee', 'A.', 'Waller', 'TRRF', 'Vice', 'President', '“Introductory', 'Remarks”', 'Lee', 'A.', 'Waller,', 'TRRF', 'Vice', 'Presi-', 'dent', 'Individual', 'Interviews', 'with', 'TRRF', 'Public', 'Board', 'Members', 'and', 'Sci-', 'entific', 'Advisory', 'Council', 'Mem-', 'bers', 'Conducted', 'by', 'TRRF', 'Treasurer', 'Philip', 'G.', 'Kuehn', 'to', 'get', 'answers', 'which', 'the', 'public', 'refrigerated', 'warehousing', 'industry', 'is', 'looking', 'for.', 'Plus', 'questions', 'from', 'the', 'floor.', 'Dr.', 'Emil', 'M.', 'Mrak,', 'University', 'of', 'Cal-', 'ifornia,', 'Chairman,', 'TRRF', 'Board;', 'Sam', 'R.', 'Cecil,', 'University', 'of', 'Georgia', 'College', 'of', 'Agriculture;', 'Dr.', 'Stanley', 'Charm,', 'Tufts', 'University', 'School', 'of', 'Medicine;', 'Dr.', 'Robert', 'H.', 'Cotton,', 'ITT', 'Continental', 'Baking', 'Company;', 'Dr.', 'Owen', 'Fennema,', 'University', 'of', 'Wis-', 'consin;', 'Dr.', 'Robert', 'E.', 'Hardenburg,', 'USDA.', 'Questions', 'and', 'Answers', 'Exhibits', 'Open', 'Capt.', 'Jack', 'Stoney', 'Room', 'TRRF', 'Scientific', 'Advisory', 'Council', 'Meeting', 'Ballroom', 'Foyer']] # noqa: E231 UpperCAmelCase__ = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words ,lowerCamelCase__ ) self.assertListEqual(encoding.boxes ,lowerCamelCase__ ) # with apply_OCR = False UpperCAmelCase__ = LayoutLMvaImageProcessor(apply_ocr=lowerCamelCase__ ) UpperCAmelCase__ = image_processing(lowerCamelCase__ ,return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 224, 224) )
716
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): return x if y == 0 else greatest_common_divisor(lowerCamelCase , x % y ) def a_ ( lowerCamelCase , lowerCamelCase ): return (x * y) // greatest_common_divisor(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase = 2_0 ): UpperCAmelCase__ = 1 for i in range(1 , n + 1 ): UpperCAmelCase__ = lcm(lowerCamelCase , lowerCamelCase ) return g if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" from copy import deepcopy class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[int] | None = None ,lowerCamelCase__ : int | None = None ): if arr is None and size is not None: UpperCAmelCase__ = size UpperCAmelCase__ = [0] * size elif arr is not None: self.init(lowerCamelCase__ ) else: raise ValueError('Either arr or size must be specified' ) def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = len(lowerCamelCase__ ) UpperCAmelCase__ = deepcopy(lowerCamelCase__ ) for i in range(1 ,self.size ): UpperCAmelCase__ = self.next_(lowerCamelCase__ ) if j < self.size: self.tree[j] += self.tree[i] def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.tree[:] for i in range(self.size - 1 ,0 ,-1 ): UpperCAmelCase__ = self.next_(lowerCamelCase__ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : int ): return index + (index & (-index)) @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : int ): return index - (index & (-index)) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value UpperCAmelCase__ = self.next_(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): self.add(lowerCamelCase__ ,value - self.get(lowerCamelCase__ ) ) def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ): if right == 0: return 0 UpperCAmelCase__ = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] UpperCAmelCase__ = self.prev(lowerCamelCase__ ) return result def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): return self.prefix(lowerCamelCase__ ) - self.prefix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : int ): return self.query(lowerCamelCase__ ,index + 1 ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : int ): value -= self.tree[0] if value < 0: return -1 UpperCAmelCase__ = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 UpperCAmelCase__ = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
717
"""simple docstring""" import warnings from functools import wraps from typing import Callable def a_ ( lowerCamelCase ): @wraps(lowerCamelCase ) def _inner_fn(*lowerCamelCase , **lowerCamelCase ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , lowerCamelCase , ) return fn(*lowerCamelCase , **lowerCamelCase ) return _inner_fn
632
0
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = (PNDMScheduler,) snake_case__ = (("num_inference_steps", 50),) def __lowerCAmelCase ( self : List[Any] ,**lowerCamelCase__ : List[Any] ): UpperCAmelCase__ = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', } config.update(**lowerCamelCase__ ) return config def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : List[str]=0 ,**lowerCamelCase__ : str ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config(**lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals UpperCAmelCase__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class.from_pretrained(lowerCamelCase__ ) new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __lowerCAmelCase ( self : str ): pass def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : int=0 ,**lowerCamelCase__ : List[Any] ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class.from_pretrained(lowerCamelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = new_scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Any ): UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(**lowerCamelCase__ ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) UpperCAmelCase__ = 10 UpperCAmelCase__ = self.dummy_model() UpperCAmelCase__ = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase__ = model(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase__ = model(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample return sample def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = dict(self.forward_default_kwargs ) UpperCAmelCase__ = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample if num_inference_steps is not None and hasattr(lowerCamelCase__ ,'set_timesteps' ): scheduler.set_timesteps(lowerCamelCase__ ) elif num_inference_steps is not None and not hasattr(lowerCamelCase__ ,'set_timesteps' ): UpperCAmelCase__ = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] UpperCAmelCase__ = dummy_past_residuals[:] UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,0 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,1 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,0 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample UpperCAmelCase__ = scheduler.step_plms(lowerCamelCase__ ,1 ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def __lowerCAmelCase ( self : List[str] ): for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowerCamelCase__ ) UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps ,torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) ,) def __lowerCAmelCase ( self : List[Any] ): for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1] ,[0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=lowerCamelCase__ ,beta_end=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): for t in [1, 5, 10]: self.check_over_forward(time_step=lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): for t, num_inference_steps in zip([1, 5, 10] ,[10, 50, 100] ): self.check_over_forward(num_inference_steps=lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase__ = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase__ = self.dummy_sample UpperCAmelCase__ = 0.1 * sample UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase__ = scheduler.step_prk(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample def __lowerCAmelCase ( self : Union[str, Any] ): with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**lowerCamelCase__ ) scheduler.step_plms(self.dummy_sample ,1 ,self.dummy_sample ).prev_sample def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = self.full_loop() UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 198.1_318 ) < 1e-2 assert abs(result_mean.item() - 0.2_5_8_0 ) < 1e-3 def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.full_loop(prediction_type='v_prediction' ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 67.3_986 ) < 1e-2 assert abs(result_mean.item() - 0.0_8_7_8 ) < 1e-3 def __lowerCAmelCase ( self : int ): # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase__ = self.full_loop(set_alpha_to_one=lowerCamelCase__ ,beta_start=0.0_1 ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 230.0_399 ) < 1e-2 assert abs(result_mean.item() - 0.2_9_9_5 ) < 1e-3 def __lowerCAmelCase ( self : Dict ): # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase__ = self.full_loop(set_alpha_to_one=lowerCamelCase__ ,beta_start=0.0_1 ) UpperCAmelCase__ = torch.sum(torch.abs(lowerCamelCase__ ) ) UpperCAmelCase__ = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 186.9_482 ) < 1e-2 assert abs(result_mean.item() - 0.2_4_3_4 ) < 1e-3
718
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCAmelCase__ : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCAmelCase__ : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowerCAmelCase__ : set[int] = {ord(char) for char in VALID_CHARS} lowerCAmelCase__ : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = "" UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 for keychar, cipherchar in zip(cycle(lowerCamelCase ) , lowerCamelCase ): UpperCAmelCase__ = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCamelCase ) return decoded def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] for key in product(lowerCamelCase , repeat=3 ): UpperCAmelCase__ = try_key(lowerCamelCase , lowerCamelCase ) if encoded is not None: possibles.append(lowerCamelCase ) return possibles def a_ ( lowerCamelCase , lowerCamelCase ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCamelCase = "p059_cipher.txt" ): UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = Path(lowerCamelCase ).parent.joinpath(lowerCamelCase ).read_text(encoding='utf-8' ) UpperCAmelCase__ = [int(lowerCamelCase ) for number in data.strip().split(',' )] UpperCAmelCase__ = filter_valid_chars(lowerCamelCase ) for common_word in COMMON_WORDS: UpperCAmelCase__ = filter_common_word(lowerCamelCase , lowerCamelCase ) if len(lowerCamelCase ) == 1: break UpperCAmelCase__ = possibles[0] return sum(ord(lowerCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
632
0
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class snake_case : """simple docstring""" snake_case__ = None def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) UpperCAmelCase__ = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] ,lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = os.path.join(lowerCamelCase__ ,'feat_extract.json' ) feat_extract_first.to_json_file(lowerCamelCase__ ) UpperCAmelCase__ = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) self.assertEqual(feat_extract_second.to_dict() ,feat_extract_first.to_dict() ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) UpperCAmelCase__ = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) self.assertEqual(feat_extract_second.to_dict() ,feat_extract_first.to_dict() ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.feature_extraction_class() self.assertIsNotNone(lowerCamelCase__ )
719
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = XLMProphetNetTokenizer snake_case__ = False snake_case__ = True def __lowerCAmelCase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = '[PAD]' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'[PAD]' ) self.assertEqual(vocab_keys[1] ,'[CLS]' ) self.assertEqual(vocab_keys[-1] ,'j' ) self.assertEqual(len(lowerCamelCase__ ) ,1_012 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_012 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) UpperCAmelCase__ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] ,) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] ,) @cached_property def __lowerCAmelCase ( self : Dict ): return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'Hello World!' UpperCAmelCase__ = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/xprophetnet-large-wiki100-cased' ,revision='1acad1643ddd54a44df6a1b797ada8373685d90e' ,)
632
0
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ : Union[str, Any] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained(lowerCamelCase__ ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = generator.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'cyberpunk 2077' UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = 'A painting of a squirrel eating a burger ' UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.text_to_image( prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = pipe.image_variation(lowerCamelCase__ ,generator=lowerCamelCase__ ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
720
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def a_ ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def a_ ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] lowerCAmelCase__ : str = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') lowerCAmelCase__ : str = sorted({word.strip().lower() for word in data.splitlines()}) lowerCAmelCase__ : Optional[Any] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": lowerCAmelCase__ : int = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
632
0
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = len(lowerCamelCase ) print('The following activities are selected:' ) # The first activity is always selected UpperCAmelCase__ = 0 print(lowerCamelCase , end=',' ) # Consider rest of the activities for j in range(lowerCamelCase ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(lowerCamelCase , end=',' ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ : Optional[int] = [1, 3, 0, 5, 8, 5] lowerCAmelCase__ : int = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
721
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class snake_case ( ctypes.Structure ): """simple docstring""" snake_case__ = [("size", ctypes.c_int), ("visible", ctypes.c_byte)] def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def a_ ( ): try: hide_cursor() yield finally: show_cursor()
632
0
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. lowerCAmelCase__ : str = 10 def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): for i in range(lowerCamelCase , lowerCamelCase ): if array[i] == target: return i return -1 def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = 0 UpperCAmelCase__ = len(lowerCamelCase ) while left <= right: if right - left < precision: return lin_search(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = (left + right) // 3 + 1 UpperCAmelCase__ = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: UpperCAmelCase__ = one_third - 1 elif array[two_third] < target: UpperCAmelCase__ = two_third + 1 else: UpperCAmelCase__ = one_third + 1 UpperCAmelCase__ = two_third - 1 else: return -1 def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if left < right: if right - left < precision: return lin_search(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = (left + right) // 3 + 1 UpperCAmelCase__ = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(lowerCamelCase , one_third - 1 , lowerCamelCase , lowerCamelCase ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , lowerCamelCase , lowerCamelCase ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ : Union[str, Any] = input('Enter numbers separated by comma:\n').strip() lowerCAmelCase__ : int = [int(item.strip()) for item in user_input.split(',')] assert collection == sorted(collection), F"List must be ordered.\n{collection}." lowerCAmelCase__ : List[Any] = int(input('Enter the number to be found in the list:\n').strip()) lowerCAmelCase__ : str = ite_ternary_search(collection, target) lowerCAmelCase__ : Any = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(F"""Iterative search: {target} found at positions: {resulta}""") print(F"""Recursive search: {target} found at positions: {resulta}""") else: print('Not found')
700
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class snake_case ( __UpperCAmelCase ): """simple docstring""" def __get__( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) UpperCAmelCase__ = '__cached_' + self.fget.__name__ UpperCAmelCase__ = getattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) if cached is None: UpperCAmelCase__ = self.fget(lowerCamelCase__ ) setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return cached def a_ ( lowerCamelCase ): UpperCAmelCase__ = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def a_ ( lowerCamelCase ): if is_torch_fx_proxy(lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return _is_numpy(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.Tensor ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.device ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_device(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch if isinstance(lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) else: return False return isinstance(lowerCamelCase , torch.dtype ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_dtype(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf return isinstance(lowerCamelCase , tf.Tensor ) def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tensorflow(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowerCamelCase , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowerCamelCase ) return type(lowerCamelCase ) == tf.Tensor def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowerCamelCase ) def a_ ( lowerCamelCase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowerCamelCase , jnp.ndarray ) def a_ ( lowerCamelCase ): return False if not is_flax_available() else _is_jax(lowerCamelCase ) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return [to_py_obj(lowerCamelCase ) for o in obj] elif is_tf_tensor(lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ).tolist() elif isinstance(lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return np.array(lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): return obj.numpy() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ) else: return obj class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = fields(self ) # Safety and consistency checks if not len(lowerCamelCase__ ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) UpperCAmelCase__ = getattr(self ,class_fields[0].name ) UpperCAmelCase__ = all(getattr(self ,field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase__ ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = first_field.items() UpperCAmelCase__ = True else: try: UpperCAmelCase__ = iter(lowerCamelCase__ ) UpperCAmelCase__ = True except TypeError: UpperCAmelCase__ = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase__ ): if ( not isinstance(lowerCamelCase__ ,(list, tuple) ) or not len(lowerCamelCase__ ) == 2 or not isinstance(element[0] ,lowerCamelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute UpperCAmelCase__ = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self ,element[0] ,element[1] ) if element[1] is not None: UpperCAmelCase__ = element[1] elif first_field is not None: UpperCAmelCase__ = first_field else: for field in class_fields: UpperCAmelCase__ = getattr(self ,field.name ) if v is not None: UpperCAmelCase__ = v def __delitem__( self : List[Any] ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : Dict ): raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : int ,**lowerCamelCase__ : Optional[Any] ): raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : List[Any] ,*lowerCamelCase__ : int ,**lowerCamelCase__ : List[Any] ): raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Optional[int] ): raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[Any] ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __setitem__( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): # Will raise a KeyException if needed super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): return tuple(self[k] for k in self.keys() ) class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : str ,lowerCamelCase__ : Optional[int] ): raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "longest" snake_case__ = "max_length" snake_case__ = "do_not_pad" class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "pt" snake_case__ = "tf" snake_case__ = "np" snake_case__ = "jax" class snake_case : """simple docstring""" def __init__( self : int ,lowerCamelCase__ : List[ContextManager] ): UpperCAmelCase__ = context_managers UpperCAmelCase__ = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase__ ) def __exit__( self : List[Any] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Dict ): self.stack.__exit__(*lowerCamelCase__ ,**lowerCamelCase__ ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( lowerCamelCase ): UpperCAmelCase__ = model_class.__name__ UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( lowerCamelCase , lowerCamelCase = "" , lowerCamelCase = "." ): def _flatten_dict(lowerCamelCase , lowerCamelCase="" , lowerCamelCase="." ): for k, v in d.items(): UpperCAmelCase__ = str(lowerCamelCase ) + delimiter + str(lowerCamelCase ) if parent_key else k if v and isinstance(lowerCamelCase , lowerCamelCase ): yield from flatten_dict(lowerCamelCase , lowerCamelCase , delimiter=lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) @contextmanager def a_ ( lowerCamelCase , lowerCamelCase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.transpose(lowerCamelCase , axes=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.T if axes is None else array.permute(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.transpose(lowerCamelCase , perm=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.transpose(lowerCamelCase , axes=lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.reshape(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.reshape(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.reshape(lowerCamelCase , lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.reshape(lowerCamelCase , lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.squeeze(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.expand_dims(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.unsqueeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.expand_dims(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.expand_dims(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.size(lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.numel() elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.size(lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): for key, value in auto_map.items(): if isinstance(lowerCamelCase , (tuple, list) ): UpperCAmelCase__ = [f'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: UpperCAmelCase__ = f'''{repo_id}--{value}''' return auto_map def a_ ( lowerCamelCase ): for base_class in inspect.getmro(lowerCamelCase ): UpperCAmelCase__ = base_class.__module__ UpperCAmelCase__ = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
632
0
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[Any]=13 ,lowerCamelCase__ : Optional[Any]=7 ,lowerCamelCase__ : str=True ,lowerCamelCase__ : str=True ,lowerCamelCase__ : List[str]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Union[str, Any]=99 ,lowerCamelCase__ : Dict=32 ,lowerCamelCase__ : List[Any]=5 ,lowerCamelCase__ : Tuple=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : List[str]="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : int=0.1 ,lowerCamelCase__ : str=512 ,lowerCamelCase__ : Optional[int]=16 ,lowerCamelCase__ : Optional[Any]=2 ,lowerCamelCase__ : Dict=0.0_2 ,lowerCamelCase__ : List[str]=4 ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_choices def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ = None if self.use_attention_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase__ = RobertaPreLayerNormConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowerCamelCase__ ,initializer_range=self.initializer_range ,) return config, input_ids, token_type_ids, attention_mask def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = True UpperCAmelCase__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = True snake_case__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = FlaxRobertaPreLayerNormModelTester(self ) @slow def __lowerCAmelCase ( self : Union[str, Any] ): for model_class_name in self.all_model_classes: UpperCAmelCase__ = model_class_name.from_pretrained('andreasmadsen/efficient_mlm_m0.40' ,from_pt=lowerCamelCase__ ) UpperCAmelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class snake_case ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained('andreasmadsen/efficient_mlm_m0.40' ,from_pt=lowerCamelCase__ ) UpperCAmelCase__ = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ,dtype=jnp.intaa ) UpperCAmelCase__ = model(lowerCamelCase__ )[0] UpperCAmelCase__ = [1, 11, 50_265] self.assertEqual(list(output.shape ) ,lowerCamelCase__ ) # compare the actual values for a slice. UpperCAmelCase__ = np.array( [[[40.4_880, 18.0_199, -5.2_3_6_7], [-1.8_8_7_7, -4.0_8_8_5, 10.7_085], [-2.2_6_1_3, -5.6_1_1_0, 7.2_6_6_5]]] ,dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] ,lowerCamelCase__ ,atol=1e-4 ) ) @slow def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = FlaxRobertaPreLayerNormModel.from_pretrained('andreasmadsen/efficient_mlm_m0.40' ,from_pt=lowerCamelCase__ ) UpperCAmelCase__ = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ,dtype=jnp.intaa ) UpperCAmelCase__ = model(lowerCamelCase__ )[0] # compare the actual values for a slice. UpperCAmelCase__ = np.array( [[[0.0_2_0_8, -0.0_3_5_6, 0.0_2_3_7], [-0.1_5_6_9, -0.0_4_1_1, -0.2_6_2_6], [0.1_8_7_9, 0.0_1_2_5, -0.0_0_8_9]]] ,dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] ,lowerCamelCase__ ,atol=1e-4 ) )
701
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ : Union[str, Any] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained(lowerCamelCase__ ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = generator.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'cyberpunk 2077' UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = 'A painting of a squirrel eating a burger ' UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.text_to_image( prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = pipe.image_variation(lowerCamelCase__ ,generator=lowerCamelCase__ ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
632
0
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
702
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
632
0
"""simple docstring""" import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : Tuple = logging.get_logger(__name__) lowerCAmelCase__ : Tuple = { 'vocab_file': 'vocab.txt', 'merges_file': 'bpe.codes', } lowerCAmelCase__ : str = { 'vocab_file': { 'vinai/phobert-base': 'https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt', 'vinai/phobert-large': 'https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt', }, 'merges_file': { 'vinai/phobert-base': 'https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes', 'vinai/phobert-large': 'https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes', }, } lowerCAmelCase__ : Dict = { 'vinai/phobert-base': 256, 'vinai/phobert-large': 256, } def a_ ( lowerCamelCase ): UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char UpperCAmelCase__ = set(lowerCamelCase ) return pairs class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowerCamelCase__ : Any ,lowerCamelCase__ : Any ,lowerCamelCase__ : List[str]="<s>" ,lowerCamelCase__ : Optional[int]="</s>" ,lowerCamelCase__ : Any="</s>" ,lowerCamelCase__ : str="<s>" ,lowerCamelCase__ : List[str]="<unk>" ,lowerCamelCase__ : Union[str, Any]="<pad>" ,lowerCamelCase__ : Optional[Any]="<mask>" ,**lowerCamelCase__ : List[Any] ,): super().__init__( bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) UpperCAmelCase__ = vocab_file UpperCAmelCase__ = merges_file UpperCAmelCase__ = {} UpperCAmelCase__ = 0 UpperCAmelCase__ = 1 UpperCAmelCase__ = 2 UpperCAmelCase__ = 3 self.add_from_file(lowerCamelCase__ ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} with open(lowerCamelCase__ ,encoding='utf-8' ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split('\n' )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:-1] ) for merge in merges] UpperCAmelCase__ = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) UpperCAmelCase__ = {} def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] UpperCAmelCase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ,lowerCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ ,token_ids_a=lowerCamelCase__ ,already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowerCAmelCase ( self : List[str] ): return len(self.encoder ) def __lowerCAmelCase ( self : Any ): return dict(self.encoder ,**self.added_tokens_encoder ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : Any ): if token in self.cache: return self.cache[token] UpperCAmelCase__ = tuple(lowerCamelCase__ ) UpperCAmelCase__ = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) UpperCAmelCase__ = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: UpperCAmelCase__ = min(lowerCamelCase__ ,key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ ,float('inf' ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(lowerCamelCase__ ): try: UpperCAmelCase__ = word.index(lowerCamelCase__ ,lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(lowerCamelCase__ ) UpperCAmelCase__ = new_word if len(lowerCamelCase__ ) == 1: break else: UpperCAmelCase__ = get_pairs(lowerCamelCase__ ) UpperCAmelCase__ = '@@ '.join(lowerCamelCase__ ) UpperCAmelCase__ = word[:-4] UpperCAmelCase__ = word return word def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : List[str] ): UpperCAmelCase__ = [] UpperCAmelCase__ = re.findall(r'\S+\n?' ,lowerCamelCase__ ) for token in words: split_tokens.extend(list(self.bpe(lowerCamelCase__ ).split(' ' ) ) ) return split_tokens def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): return self.encoder.get(lowerCamelCase__ ,self.encoder.get(self.unk_token ) ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[Any] ): return self.decoder.get(lowerCamelCase__ ,self.unk_token ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : Tuple ): UpperCAmelCase__ = ' '.join(lowerCamelCase__ ).replace('@@ ' ,'' ).strip() return out_string def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) if os.path.abspath(self.merges_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.merges_file ,lowerCamelCase__ ) return out_vocab_file, out_merge_file def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Dict ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): try: with open(lowerCamelCase__ ,'r' ,encoding='utf-8' ) as fd: self.add_from_file(lowerCamelCase__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return UpperCAmelCase__ = f.readlines() for lineTmp in lines: UpperCAmelCase__ = lineTmp.strip() UpperCAmelCase__ = line.rfind(' ' ) if idx == -1: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt>\'' ) UpperCAmelCase__ = line[:idx] UpperCAmelCase__ = len(self.encoder )
703
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
632
0
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed lowerCAmelCase__ : Tuple = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) lowerCAmelCase__ : List[str] = 'sshleifer/student_marian_en_ro_6_1' lowerCAmelCase__ : Optional[int] = 'sshleifer/tiny-mbart' @require_torch class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Optional[int]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : int=True ,): UpperCAmelCase__ = self.run_trainer( eval_steps=1 ,max_len=12 ,model_name=lowerCamelCase__ ,num_train_epochs=1 ,distributed=lowerCamelCase__ ,extra_args_str=lowerCamelCase__ ,predict_with_generate=lowerCamelCase__ ,do_train=lowerCamelCase__ ,do_eval=lowerCamelCase__ ,do_predict=lowerCamelCase__ ,) UpperCAmelCase__ = TrainerState.load_from_json(os.path.join(lowerCamelCase__ ,'trainer_state.json' ) ).log_history if not do_eval: return UpperCAmelCase__ = [log for log in logs if 'eval_loss' in log.keys()] UpperCAmelCase__ = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase__ = eval_metrics[-1] assert isinstance(last_step_stats['eval_bleu'] ,lowerCamelCase__ ) assert not math.isnan(float(last_step_stats['eval_loss'] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def __lowerCAmelCase ( self : Any ): self.run_seqaseq_quick() @require_torch_multi_gpu def __lowerCAmelCase ( self : Tuple ): self.run_seqaseq_quick(distributed=lowerCamelCase__ ) @require_torch_multi_gpu def __lowerCAmelCase ( self : int ): self.run_seqaseq_quick(distributed=lowerCamelCase__ ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def __lowerCAmelCase ( self : Union[str, Any] ): self.run_seqaseq_quick(distributed=lowerCamelCase__ ,extra_args_str='--sharded_ddp simple' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def __lowerCAmelCase ( self : Optional[int] ): self.run_seqaseq_quick(distributed=lowerCamelCase__ ,extra_args_str='--sharded_ddp simple --fp16' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def __lowerCAmelCase ( self : Tuple ): self.run_seqaseq_quick(distributed=lowerCamelCase__ ,extra_args_str='--sharded_ddp zero_dp_2' ,predict_with_generate=lowerCamelCase__ ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def __lowerCAmelCase ( self : Dict ): self.run_seqaseq_quick( distributed=lowerCamelCase__ ,extra_args_str='--sharded_ddp zero_dp_2 --fp16' ,predict_with_generate=lowerCamelCase__ ) @require_apex @require_torch_gpu def __lowerCAmelCase ( self : int ): # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=lowerCamelCase__ ,extra_args_str='--fp16 --fp16_backend=apex' ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowerCamelCase__ ,extra_args_str='--fp16 --fp16_backend=apex' ) @parameterized.expand(['base', 'low', 'high', 'mixed'] ) @require_torch_multi_gpu def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : Any ): # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout UpperCAmelCase__ = { # test with the default log_level - should be info and thus log info once 'base': {'extra_args_str': '', 'n_matches': 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes 'low': {'extra_args_str': '--log_level debug --log_level_replica debug', 'n_matches': 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica 'high': {'extra_args_str': '--log_level error --log_level_replica debug', 'n_matches': 1}, # test with high log_level and log_level_replica - should be quiet on all processes 'mixed': {'extra_args_str': '--log_level error --log_level_replica error', 'n_matches': 0}, } UpperCAmelCase__ = experiments[experiment_id] UpperCAmelCase__ = {'distributed': True, 'predict_with_generate': False, 'do_eval': False, 'do_predict': False} UpperCAmelCase__ = 'Running training' with CaptureStderr() as cl: self.run_seqaseq_quick(**lowerCamelCase__ ,extra_args_str=data['extra_args_str'] ) UpperCAmelCase__ = len(re.findall(lowerCamelCase__ ,cl.err ) ) self.assertEqual(lowerCamelCase__ ,data['n_matches'] ) @slow def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.run_trainer( eval_steps=2 ,max_len=128 ,model_name=lowerCamelCase__ ,learning_rate=3e-4 ,num_train_epochs=10 ,distributed=lowerCamelCase__ ,) # Check metrics UpperCAmelCase__ = TrainerState.load_from_json(os.path.join(lowerCamelCase__ ,'trainer_state.json' ) ).log_history UpperCAmelCase__ = [log for log in logs if 'eval_loss' in log.keys()] UpperCAmelCase__ = eval_metrics[0] UpperCAmelCase__ = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats['eval_bleu'] ,lowerCamelCase__ ) # test if do_predict saves generations and metrics UpperCAmelCase__ = os.listdir(lowerCamelCase__ ) UpperCAmelCase__ = {os.path.basename(lowerCamelCase__ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def __lowerCAmelCase ( self : Tuple ): from transformers.training_args import OptimizerNames def train_and_return_metrics(lowerCamelCase__ : str ) -> Tuple[int, float]: UpperCAmelCase__ = '--skip_memory_metrics 0' UpperCAmelCase__ = self.run_trainer( max_len=128 ,model_name=lowerCamelCase__ ,learning_rate=3e-4 ,num_train_epochs=1 ,optim=lowerCamelCase__ ,distributed=lowerCamelCase__ ,extra_args_str=lowerCamelCase__ ,do_eval=lowerCamelCase__ ,do_predict=lowerCamelCase__ ,n_gpus_to_use=1 ,) # Check metrics UpperCAmelCase__ = TrainerState.load_from_json(Path(lowerCamelCase__ ,'trainer_state.json' ) ).log_history UpperCAmelCase__ = int(logs[0]['train_mem_gpu_peaked_delta'] / 2**20 ) UpperCAmelCase__ = int(logs[0]['train_mem_gpu_alloc_delta'] / 2**20 ) UpperCAmelCase__ = logs[0]['train_loss'] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase__ = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase__ = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase__ = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase__ = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase__ = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowerCamelCase__ ,lowerCamelCase__ ,'should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got' f''' a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and''' f''' gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB''' ,) self.assertGreater( lowerCamelCase__ ,lowerCamelCase__ ,'should use ~150MB less total gpu memory with BNB, compared to without it for this model but got' f''' a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and''' f''' gpu_total_mem_bnb={gpu_total_mem_bnb}MB''' ,) self.assertEqual( lowerCamelCase__ ,lowerCamelCase__ ,f'''loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}''' ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : str ,lowerCamelCase__ : int ,lowerCamelCase__ : float = 3e-3 ,lowerCamelCase__ : str = "adafactor" ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : str = None ,lowerCamelCase__ : int = 0 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : int = None ,): UpperCAmelCase__ = self.test_file_dir / '../fixtures/tests_samples/wmt_en_ro' UpperCAmelCase__ = self.get_auto_remove_tmp_dir() UpperCAmelCase__ = f''' --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(lowerCamelCase__ )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(lowerCamelCase__ )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX '''.split() UpperCAmelCase__ = f''' --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(lowerCamelCase__ )} '''.split() UpperCAmelCase__ = '\n --do_predict\n '.split() UpperCAmelCase__ = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f'''--optim {optim}'''.split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase__ = get_gpu_count() UpperCAmelCase__ = get_torch_dist_unique_port() UpperCAmelCase__ = f''' -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py '''.split() UpperCAmelCase__ = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowerCamelCase__ ,env=self.get_env() ) else: UpperCAmelCase__ = ['run_translation.py'] + args with patch.object(lowerCamelCase__ ,'argv' ,lowerCamelCase__ ): main() return output_dir
704
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
632
0
"""simple docstring""" def a_ ( lowerCamelCase ): UpperCAmelCase__ = current_set.copy() for row_index, row in enumerate(lowerCamelCase ): UpperCAmelCase__ = row[0] for column_index, column in enumerate(lowerCamelCase ): if magnitude == 0: UpperCAmelCase__ = column continue UpperCAmelCase__ = column / magnitude # Subtract to cancel term UpperCAmelCase__ = current_set[0] UpperCAmelCase__ = [first_row] UpperCAmelCase__ = current_set[1::] for row in current_set: UpperCAmelCase__ = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(lowerCamelCase ) continue for column_index in range(len(lowerCamelCase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(lowerCamelCase ) # Create next recursion iteration set if len(final_set[0] ) != 3: UpperCAmelCase__ = final_set[0] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) UpperCAmelCase__ = simplify(lowerCamelCase ) for i in range(len(lowerCamelCase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , lowerCamelCase ) UpperCAmelCase__ = resultant return final_set def a_ ( lowerCamelCase ): if len(lowerCamelCase ) == 0: raise IndexError('solve_simultaneous() requires n lists of length n+1' ) UpperCAmelCase__ = len(lowerCamelCase ) + 1 if any(len(lowerCamelCase ) != _length for item in equations ): raise IndexError('solve_simultaneous() requires n lists of length n+1' ) for row in equations: if any(not isinstance(lowerCamelCase , (int, float) ) for column in row ): raise ValueError('solve_simultaneous() requires lists of integers' ) if len(lowerCamelCase ) == 1: return [equations[0][-1] / equations[0][0]] UpperCAmelCase__ = equations.copy() if any(0 in row for row in data_set ): UpperCAmelCase__ = data_set.copy() UpperCAmelCase__ = [] for row_index, row in enumerate(lowerCamelCase ): if 0 not in row: UpperCAmelCase__ = data_set.pop(lowerCamelCase ) break if not full_row: raise ValueError('solve_simultaneous() requires at least 1 full equation' ) data_set.insert(0 , lowerCamelCase ) UpperCAmelCase__ = data_set.copy() UpperCAmelCase__ = simplify(lowerCamelCase ) UpperCAmelCase__ = simplified[::-1] UpperCAmelCase__ = [] for row in simplified: UpperCAmelCase__ = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue UpperCAmelCase__ = row.copy()[: len(lowerCamelCase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(lowerCamelCase ) == 0: solutions.append(0 ) continue UpperCAmelCase__ = temp_row[1::] UpperCAmelCase__ = temp_row[::-1] for column_index, column in enumerate(lowerCamelCase ): current_solution -= column * solutions[column_index] solutions.append(lowerCamelCase ) UpperCAmelCase__ = [] for item in solutions: final.append(float(round(lowerCamelCase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ : Dict = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
705
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
632
0
"""simple docstring""" import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets lowerCAmelCase__ : str = '\\n@inproceedings{snover-etal-2006-study,\n title = "A Study of Translation Edit Rate with Targeted Human Annotation",\n author = "Snover, Matthew and\n Dorr, Bonnie and\n Schwartz, Rich and\n Micciulla, Linnea and\n Makhoul, John",\n booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers",\n month = aug # " 8-12",\n year = "2006",\n address = "Cambridge, Massachusetts, USA",\n publisher = "Association for Machine Translation in the Americas",\n url = "https://aclanthology.org/2006.amta-papers.25",\n pages = "223--231",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' lowerCAmelCase__ : Union[str, Any] = '\\nTER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a\nhypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu\n(https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found\nhere: https://github.com/jhclark/tercom.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information.\n' lowerCAmelCase__ : Optional[Any] = '\nProduces TER scores alongside the number of edits and reference length.\n\nArgs:\n predictions (list of str): The system stream (a sequence of segments).\n references (list of list of str): A list of one or more reference streams (each a sequence of segments).\n normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters,\n as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana.\n Only applies if `normalized = True`. Defaults to `False`.\n case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`.\n\nReturns:\n \'score\' (float): TER score (num_edits / sum_ref_lengths * 100)\n \'num_edits\' (int): The cumulative number of edits\n \'ref_length\' (float): The cumulative average reference length\n\nExamples:\n Example 1:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0}\n\n Example 2:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0}\n\n Example 3:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... normalized=True,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5}\n\n Example 4:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0}\n\n Example 5:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): """simple docstring""" def __lowerCAmelCase ( self : Union[str, Any] ): if version.parse(scb.__version__ ) < version.parse('1.4.12' ): raise ImportWarning( 'To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n' 'You can install it with `pip install "sacrebleu>=1.4.12"`.' ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,homepage='http://www.cs.umd.edu/~snover/tercom/' ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { 'predictions': datasets.Value('string' ,id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' ,id='sequence' ) ,id='references' ), } ) ,codebase_urls=['https://github.com/mjpost/sacreBLEU#ter'] ,reference_urls=[ 'https://github.com/jhclark/tercom', ] ,) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): UpperCAmelCase__ = len(references[0] ) if any(len(lowerCamelCase__ ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) UpperCAmelCase__ = [[refs[i] for refs in references] for i in range(lowerCamelCase__ )] UpperCAmelCase__ = TER( normalized=lowerCamelCase__ ,no_punct=lowerCamelCase__ ,asian_support=lowerCamelCase__ ,case_sensitive=lowerCamelCase__ ,) UpperCAmelCase__ = sb_ter.corpus_score(lowerCamelCase__ ,lowerCamelCase__ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
706
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ : int = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Union[str, Any] = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
632
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCAmelCase__ : Dict = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ["pixel_values"] def __init__( self : int ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = PILImageResampling.BICUBIC ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Union[int, float] = 1 / 255 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : bool = True ,**lowerCamelCase__ : Any ,): super().__init__(**lowerCamelCase__ ) UpperCAmelCase__ = size if size is not None else {'shortest_edge': 224} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ) UpperCAmelCase__ = crop_size if crop_size is not None else {'height': 224, 'width': 224} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ,param_name='crop_size' ) UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = resample UpperCAmelCase__ = do_center_crop UpperCAmelCase__ = crop_size UpperCAmelCase__ = do_rescale UpperCAmelCase__ = rescale_factor UpperCAmelCase__ = do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCAmelCase__ = image_std if image_std is not None else OPENAI_CLIP_STD UpperCAmelCase__ = do_convert_rgb def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : PILImageResampling = PILImageResampling.BICUBIC ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : str ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) UpperCAmelCase__ = get_resize_output_image_size(lowerCamelCase__ ,size=size['shortest_edge'] ,default_to_square=lowerCamelCase__ ) return resize(lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Optional[int] ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(lowerCamelCase__ ,size=(size['height'], size['width']) ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[int, float] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : List[Any] ,): return rescale(lowerCamelCase__ ,scale=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Optional[Any] ,): return normalize(lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : ImageInput ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : int = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : float = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,lowerCamelCase__ : Optional[ChannelDimension] = ChannelDimension.FIRST ,**lowerCamelCase__ : Optional[int] ,): UpperCAmelCase__ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ = size if size is not None else self.size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='size' ,default_to_square=lowerCamelCase__ ) UpperCAmelCase__ = resample if resample is not None else self.resample UpperCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='crop_size' ,default_to_square=lowerCamelCase__ ) UpperCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase__ = image_std if image_std is not None else self.image_std UpperCAmelCase__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCAmelCase__ = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCAmelCase__ = [convert_to_rgb(lowerCamelCase__ ) for image in images] # All transformations expect numpy arrays. UpperCAmelCase__ = [to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: UpperCAmelCase__ = [self.resize(image=lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ) for image in images] if do_center_crop: UpperCAmelCase__ = [self.center_crop(image=lowerCamelCase__ ,size=lowerCamelCase__ ) for image in images] if do_rescale: UpperCAmelCase__ = [self.rescale(image=lowerCamelCase__ ,scale=lowerCamelCase__ ) for image in images] if do_normalize: UpperCAmelCase__ = [self.normalize(image=lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ) for image in images] UpperCAmelCase__ = [to_channel_dimension_format(lowerCamelCase__ ,lowerCamelCase__ ) for image in images] UpperCAmelCase__ = {'pixel_values': images} return BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ )
707
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase__ : Optional[int] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( image=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
632
0
"""simple docstring""" import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) lowerCAmelCase__ : Any = { 'iou_prediction_head.layers.0': 'iou_prediction_head.proj_in', 'iou_prediction_head.layers.1': 'iou_prediction_head.layers.0', 'iou_prediction_head.layers.2': 'iou_prediction_head.proj_out', 'mask_decoder.output_upscaling.0': 'mask_decoder.upscale_conv1', 'mask_decoder.output_upscaling.1': 'mask_decoder.upscale_layer_norm', 'mask_decoder.output_upscaling.3': 'mask_decoder.upscale_conv2', 'mask_downscaling.0': 'mask_embed.conv1', 'mask_downscaling.1': 'mask_embed.layer_norm1', 'mask_downscaling.3': 'mask_embed.conv2', 'mask_downscaling.4': 'mask_embed.layer_norm2', 'mask_downscaling.6': 'mask_embed.conv3', 'point_embeddings': 'point_embed', 'pe_layer.positional_encoding_gaussian_matrix': 'shared_embedding.positional_embedding', 'image_encoder': 'vision_encoder', 'neck.0': 'neck.conv1', 'neck.1': 'neck.layer_norm1', 'neck.2': 'neck.conv2', 'neck.3': 'neck.layer_norm2', 'patch_embed.proj': 'patch_embed.projection', '.norm': '.layer_norm', 'blocks': 'layers', } def a_ ( lowerCamelCase ): UpperCAmelCase__ = {} state_dict.pop('pixel_mean' , lowerCamelCase ) state_dict.pop('pixel_std' , lowerCamelCase ) UpperCAmelCase__ = r'.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase__ = key.replace(lowerCamelCase , lowerCamelCase ) if re.match(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = int(re.match(lowerCamelCase , lowerCamelCase ).group(2 ) ) if layer_nb == 0: UpperCAmelCase__ = key.replace('layers.0' , 'proj_in' ) elif layer_nb == 1: UpperCAmelCase__ = key.replace('layers.1' , 'layers.0' ) elif layer_nb == 2: UpperCAmelCase__ = key.replace('layers.2' , 'proj_out' ) UpperCAmelCase__ = value UpperCAmelCase__ = model_state_dict[ 'prompt_encoder.shared_embedding.positional_embedding' ] return model_state_dict def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase="ybelkada/segment-anything" ): UpperCAmelCase__ = hf_hub_download(lowerCamelCase , f'''checkpoints/{model_name}.pth''' ) if "sam_vit_b" in model_name: UpperCAmelCase__ = SamConfig() elif "sam_vit_l" in model_name: UpperCAmelCase__ = SamVisionConfig( hidden_size=1_0_2_4 , num_hidden_layers=2_4 , num_attention_heads=1_6 , global_attn_indexes=[5, 1_1, 1_7, 2_3] , ) UpperCAmelCase__ = SamConfig( vision_config=lowerCamelCase , ) elif "sam_vit_h" in model_name: UpperCAmelCase__ = SamVisionConfig( hidden_size=1_2_8_0 , num_hidden_layers=3_2 , num_attention_heads=1_6 , global_attn_indexes=[7, 1_5, 2_3, 3_1] , ) UpperCAmelCase__ = SamConfig( vision_config=lowerCamelCase , ) UpperCAmelCase__ = torch.load(lowerCamelCase , map_location='cpu' ) UpperCAmelCase__ = replace_keys(lowerCamelCase ) UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase ) UpperCAmelCase__ = SamModel(lowerCamelCase ) hf_model.load_state_dict(lowerCamelCase ) UpperCAmelCase__ = hf_model.to('cuda' ) UpperCAmelCase__ = 'https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png' UpperCAmelCase__ = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ).convert('RGB' ) UpperCAmelCase__ = [[[4_0_0, 6_5_0]]] UpperCAmelCase__ = [[1]] UpperCAmelCase__ = processor(images=np.array(lowerCamelCase ) , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): UpperCAmelCase__ = hf_model(**lowerCamelCase ) UpperCAmelCase__ = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 UpperCAmelCase__ = processor( images=np.array(lowerCamelCase ) , input_points=lowerCamelCase , input_labels=lowerCamelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): UpperCAmelCase__ = hf_model(**lowerCamelCase ) UpperCAmelCase__ = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 UpperCAmelCase__ = ((7_5, 2_7_5, 1_7_2_5, 8_5_0),) UpperCAmelCase__ = processor(images=np.array(lowerCamelCase ) , input_boxes=lowerCamelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): UpperCAmelCase__ = hf_model(**lowerCamelCase ) UpperCAmelCase__ = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. UpperCAmelCase__ = [[[4_0_0, 6_5_0], [8_0_0, 6_5_0]]] UpperCAmelCase__ = [[1, 1]] UpperCAmelCase__ = processor( images=np.array(lowerCamelCase ) , input_points=lowerCamelCase , input_labels=lowerCamelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): UpperCAmelCase__ = hf_model(**lowerCamelCase ) UpperCAmelCase__ = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": lowerCAmelCase__ : int = argparse.ArgumentParser() lowerCAmelCase__ : Tuple = ['sam_vit_b_01ec64', 'sam_vit_h_4b8939', 'sam_vit_l_0b3195'] parser.add_argument( '--model_name', default='sam_vit_h_4b8939', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) parser.add_argument( '--model_hub_id', default='ybelkada/segment-anything', choices=choices, type=str, help='Path to hf config.json of model to convert', ) lowerCAmelCase__ : List[Any] = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
708
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def a_ ( lowerCamelCase = 2_0_0_0_0_0_0 ): UpperCAmelCase__ = [0] UpperCAmelCase__ = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target UpperCAmelCase__ = 0 # the area corresponding to the grid that gives the product closest to target UpperCAmelCase__ = 0 # an estimate of b, using the quadratic formula UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the triangle number corresponding to b_floor UpperCAmelCase__ = 42 # the triangle number corresponding to b_ceil UpperCAmelCase__ = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): UpperCAmelCase__ = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 UpperCAmelCase__ = floor(lowerCamelCase ) UpperCAmelCase__ = ceil(lowerCamelCase ) UpperCAmelCase__ = triangle_numbers[b_floor] UpperCAmelCase__ = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_first_guess * triangle_a UpperCAmelCase__ = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_second_guess * triangle_a UpperCAmelCase__ = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
632
0
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case : """simple docstring""" def __init__( self : List[str] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Any=99 ,lowerCamelCase__ : Optional[Any]=13 ,lowerCamelCase__ : str=7 ,lowerCamelCase__ : int=9 ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=False ,lowerCamelCase__ : Dict=32 ,lowerCamelCase__ : Optional[int]=5 ,lowerCamelCase__ : Optional[int]=4 ,lowerCamelCase__ : str=37 ,lowerCamelCase__ : Optional[Any]=8 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0_0_2 ,lowerCamelCase__ : Dict=1 ,lowerCamelCase__ : Optional[Any]=0 ,lowerCamelCase__ : int=0 ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = encoder_seq_length UpperCAmelCase__ = decoder_seq_length # For common tests UpperCAmelCase__ = self.decoder_seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = d_ff UpperCAmelCase__ = relative_attention_num_buckets UpperCAmelCase__ = dropout_rate UpperCAmelCase__ = initializer_factor UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = pad_token_id UpperCAmelCase__ = decoder_start_token_id UpperCAmelCase__ = None UpperCAmelCase__ = decoder_layers def __lowerCAmelCase ( self : Tuple ): return TaConfig.from_pretrained('google/umt5-base' ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Dict=None ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : int=None ,): if attention_mask is None: UpperCAmelCase__ = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: UpperCAmelCase__ = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: UpperCAmelCase__ = torch.ones(config.num_hidden_layers ,config.num_attention_heads ,device=lowerCamelCase__ ) if decoder_head_mask is None: UpperCAmelCase__ = torch.ones(config.num_decoder_layers ,config.num_attention_heads ,device=lowerCamelCase__ ) if cross_attn_head_mask is None: UpperCAmelCase__ = torch.ones( config.num_decoder_layers ,config.num_attention_heads ,device=lowerCamelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = ids_tensor([self.batch_size, self.encoder_seq_length] ,self.vocab_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.decoder_seq_length] ,self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input UpperCAmelCase__ = input_ids.clamp(self.pad_token_id + 1 ) UpperCAmelCase__ = decoder_input_ids.clamp(self.pad_token_id + 1 ) UpperCAmelCase__ = self.get_config() UpperCAmelCase__ = config.num_attention_heads UpperCAmelCase__ = self.prepare_inputs_dict(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return config, input_dict def __lowerCAmelCase ( self : int ): UpperCAmelCase__ , UpperCAmelCase__ = self.prepare_config_and_inputs() return config, inputs_dict def __lowerCAmelCase ( self : str ): return TaConfig( vocab_size=166 ,d_model=self.hidden_size ,d_ff=self.d_ff ,d_kv=self.hidden_size // self.num_attention_heads ,num_layers=self.num_hidden_layers ,num_decoder_layers=self.decoder_layers ,num_heads=self.num_attention_heads ,relative_attention_num_buckets=self.relative_attention_num_buckets ,dropout_rate=self.dropout_rate ,initializer_factor=self.initializer_factor ,eos_token_id=self.eos_token_id ,bos_token_id=self.pad_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.decoder_start_token_id ,) def __lowerCAmelCase ( self : str ): return TaConfig( vocab_size=self.vocab_size ,d_model=self.hidden_size ,d_ff=self.d_ff ,d_kv=self.hidden_size // self.num_attention_heads ,num_layers=self.num_hidden_layers ,num_decoder_layers=self.decoder_layers ,num_heads=self.num_attention_heads ,relative_attention_num_buckets=self.relative_attention_num_buckets ,dropout_rate=self.dropout_rate ,initializer_factor=self.initializer_factor ,eos_token_id=self.eos_token_id ,bos_token_id=self.pad_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.decoder_start_token_id ,) def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Any ,lowerCamelCase__ : str ,): UpperCAmelCase__ = UMTaModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCAmelCase__ = model( input_ids=lowerCamelCase__ ,decoder_input_ids=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,decoder_attention_mask=lowerCamelCase__ ,) UpperCAmelCase__ = model(input_ids=lowerCamelCase__ ,decoder_input_ids=lowerCamelCase__ ) UpperCAmelCase__ = result.last_hidden_state UpperCAmelCase__ = result.past_key_values UpperCAmelCase__ = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() ,(self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() ,(self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase__ ) ,config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) ,4 ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any ,): UpperCAmelCase__ = UMTaModel(config=lowerCamelCase__ ).get_decoder().to(lowerCamelCase__ ).eval() # first forward pass UpperCAmelCase__ = model(lowerCamelCase__ ,use_cache=lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ,use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) UpperCAmelCase__ , UpperCAmelCase__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase__ = ids_tensor((self.batch_size, 1) ,config.vocab_size ) # append to next input_ids and UpperCAmelCase__ = torch.cat([input_ids, next_tokens] ,dim=-1 ) UpperCAmelCase__ = model(lowerCamelCase__ )['last_hidden_state'] UpperCAmelCase__ = model(lowerCamelCase__ ,past_key_values=lowerCamelCase__ )['last_hidden_state'] # select random slice UpperCAmelCase__ = ids_tensor((1,) ,output_from_past.shape[-1] ).item() UpperCAmelCase__ = output_from_no_past[:, -1, random_slice_idx].detach() UpperCAmelCase__ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1e-3 ) ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : str ,): UpperCAmelCase__ = UMTaModel(config=lowerCamelCase__ ).to(lowerCamelCase__ ).half().eval() UpperCAmelCase__ = model(**lowerCamelCase__ )['last_hidden_state'] self.parent.assertFalse(torch.isnan(lowerCamelCase__ ).any().item() ) @require_torch class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) snake_case__ = (UMTaForConditionalGeneration,) if is_torch_available() else () snake_case__ = ( { "conversational": UMTaForConditionalGeneration, "feature-extraction": UMTaModel, "summarization": UMTaForConditionalGeneration, "text2text-generation": UMTaForConditionalGeneration, "translation": UMTaForConditionalGeneration, "question-answering": UMTaForQuestionAnswering, } if is_torch_available() else {} ) snake_case__ = True snake_case__ = False snake_case__ = False snake_case__ = True snake_case__ = True # The small UMT5 model needs higher percentages for CPU/MP tests snake_case__ = [0.8, 0.9] def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase__ ,(config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) ,f'''{tmpdirname}/t5_test.onnx''' ,export_params=lowerCamelCase__ ,opset_version=9 ,input_names=['input_ids', 'decoder_input_ids'] ,) @unittest.skipIf(torch_device == 'cpu' ,'Cant do half precision' ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase__ ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ = config_and_inputs[0] UpperCAmelCase__ = UMTaForConditionalGeneration(lowerCamelCase__ ).eval() model.to(lowerCamelCase__ ) UpperCAmelCase__ = { 'head_mask': torch.zeros(config.num_layers ,config.num_heads ,device=lowerCamelCase__ ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers ,config.num_heads ,device=lowerCamelCase__ ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers ,config.num_heads ,device=lowerCamelCase__ ), } for attn_name, (name, mask) in zip(lowerCamelCase__ ,head_masking.items() ): UpperCAmelCase__ = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": UpperCAmelCase__ = torch.ones( config.num_decoder_layers ,config.num_heads ,device=lowerCamelCase__ ) UpperCAmelCase__ = model.generate( config_and_inputs[1]['input_ids'] ,num_beams=1 ,max_length=3 ,output_attentions=lowerCamelCase__ ,return_dict_in_generate=lowerCamelCase__ ,**lowerCamelCase__ ,) # We check the state of decoder_attentions and cross_attentions just from the last step UpperCAmelCase__ = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) ,0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def __lowerCAmelCase ( self : List[str] ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case ( unittest.TestCase ): """simple docstring""" @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' ,return_dict=lowerCamelCase__ ).to(lowerCamelCase__ ) UpperCAmelCase__ = AutoTokenizer.from_pretrained('google/umt5-small' ,use_fast=lowerCamelCase__ ,legacy=lowerCamelCase__ ) UpperCAmelCase__ = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] UpperCAmelCase__ = tokenizer(lowerCamelCase__ ,return_tensors='pt' ,padding=lowerCamelCase__ ).input_ids # fmt: off UpperCAmelCase__ = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = model.generate(input_ids.to(lowerCamelCase__ ) ) UpperCAmelCase__ = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] UpperCAmelCase__ = tokenizer.batch_decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ )
709
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : Optional[Any] = ['model.decoder.embed_positions.weights'] def a_ ( lowerCamelCase ): if "emb" in name: UpperCAmelCase__ = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: UpperCAmelCase__ = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: UpperCAmelCase__ = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: UpperCAmelCase__ = name.replace('linear1' , 'fc1' ) if "linear2" in name: UpperCAmelCase__ = name.replace('linear2' , 'fc2' ) if "norm1" in name: UpperCAmelCase__ = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: UpperCAmelCase__ = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: UpperCAmelCase__ = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: UpperCAmelCase__ = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: UpperCAmelCase__ = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase__ = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = list(state_dict.keys() ) UpperCAmelCase__ = {} for key in keys: UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) UpperCAmelCase__ = rename_keys(lowerCamelCase ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase__ = val[:hidden_size, :] UpperCAmelCase__ = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase__ = val else: UpperCAmelCase__ = val return state_dict, enc_dec_proj_state_dict def a_ ( lowerCamelCase ): if checkpoint == "small": # default config values UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif checkpoint == "medium": UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 2_4 elif checkpoint == "large": UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 3_2 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) UpperCAmelCase__ = MusicgenDecoderConfig( hidden_size=lowerCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=lowerCamelCase , num_attention_heads=lowerCamelCase , ) return config @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="cpu" ): UpperCAmelCase__ = MusicGen.get_pretrained(lowerCamelCase , device=lowerCamelCase ) UpperCAmelCase__ = decoder_config_from_checkpoint(lowerCamelCase ) UpperCAmelCase__ = fairseq_model.lm.state_dict() UpperCAmelCase__ , UpperCAmelCase__ = rename_state_dict( lowerCamelCase , hidden_size=decoder_config.hidden_size ) UpperCAmelCase__ = TaEncoderModel.from_pretrained('t5-base' ) UpperCAmelCase__ = EncodecModel.from_pretrained('facebook/encodec_32khz' ) UpperCAmelCase__ = MusicgenForCausalLM(lowerCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase__ , UpperCAmelCase__ = decoder.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model UpperCAmelCase__ = MusicgenForConditionalGeneration(text_encoder=lowerCamelCase , audio_encoder=lowerCamelCase , decoder=lowerCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(lowerCamelCase ) # check we can do a forward pass UpperCAmelCase__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase__ = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits if logits.shape != (8, 1, 2_0_4_8): raise ValueError('Incorrect shape for logits' ) # now construct the processor UpperCAmelCase__ = AutoTokenizer.from_pretrained('t5-base' ) UpperCAmelCase__ = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) UpperCAmelCase__ = MusicgenProcessor(feature_extractor=lowerCamelCase , tokenizer=lowerCamelCase ) # set the appropriate bos/pad token ids UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 2_0_4_8 # set other default generation config params UpperCAmelCase__ = int(3_0 * audio_encoder.config.frame_rate ) UpperCAmelCase__ = True UpperCAmelCase__ = 3.0 if pytorch_dump_folder is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(lowerCamelCase ) processor.save_pretrained(lowerCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(lowerCamelCase ) processor.push_to_hub(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
632
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ : int = { 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : int = ['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = ['LayoutLMv2FeatureExtractor'] lowerCAmelCase__ : Union[str, Any] = ['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys lowerCAmelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
710
"""simple docstring""" lowerCAmelCase__ : Tuple = range(2, 20 + 1) lowerCAmelCase__ : Optional[Any] = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase__ : dict[int, dict[int, list[list[int]]]] = {} def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(a_i[j] for j in range(lowerCamelCase , len(lowerCamelCase ) ) ) UpperCAmelCase__ = sum(a_i[j] * base[j] for j in range(min(len(lowerCamelCase ) , lowerCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ = 0, 0 UpperCAmelCase__ = n - i UpperCAmelCase__ = memo.get(lowerCamelCase ) if sub_memo is not None: UpperCAmelCase__ = sub_memo.get(lowerCamelCase ) if jumps is not None and len(lowerCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ = -1 for _k in range(len(lowerCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ = diff + c for j in range(min(lowerCamelCase , len(lowerCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) if new_c > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: UpperCAmelCase__ = [] else: UpperCAmelCase__ = {c: []} UpperCAmelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , k - 1 , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ = compute(lowerCamelCase , lowerCamelCase , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ = 0 while j < len(lowerCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowerCamelCase , (diff, dn, k) ) return (diff, dn) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if i >= n: return 0, i if k > len(lowerCamelCase ): a_i.extend([0 for _ in range(k - len(lowerCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 0, 0, 0 for j in range(len(lowerCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ = ds_c + ds_b diff += addend UpperCAmelCase__ = 0 for j in range(lowerCamelCase ): UpperCAmelCase__ = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return diff, i - start_i def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): for j in range(lowerCamelCase , len(lowerCamelCase ) ): UpperCAmelCase__ = digits[j] + addend if s >= 1_0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) UpperCAmelCase__ = addend // 1_0 + quotient else: UpperCAmelCase__ = s UpperCAmelCase__ = addend // 1_0 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) digits.append(lowerCamelCase ) def a_ ( lowerCamelCase = 1_0**1_5 ): UpperCAmelCase__ = [1] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , 2_0 , i + dn , lowerCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ = 0 for j in range(len(lowerCamelCase ) ): a_n += digits[j] * 1_0**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def a_ ( lowerCamelCase , lowerCamelCase=7 ): UpperCAmelCase__ = None if token is not None: UpperCAmelCase__ = {'Accept': 'application/vnd.github+json', 'Authorization': f'''Bearer {token}'''} # The id of a workflow (not of a workflow run) UpperCAmelCase__ = '636036' UpperCAmelCase__ = f'''https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs''' # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += f'''?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}''' UpperCAmelCase__ = requests.get(lowerCamelCase , headers=lowerCamelCase ).json() return result["workflow_runs"] def a_ ( lowerCamelCase ): UpperCAmelCase__ = get_daily_ci_runs(lowerCamelCase ) UpperCAmelCase__ = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": UpperCAmelCase__ = workflow_run['id'] break return workflow_run_id def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = get_last_daily_ci_runs(lowerCamelCase ) if workflow_run_id is not None: UpperCAmelCase__ = get_artifacts_links(worflow_run_id=lowerCamelCase , token=lowerCamelCase ) for artifact_name in artifact_names: if artifact_name in artifacts_links: UpperCAmelCase__ = artifacts_links[artifact_name] download_artifact( artifact_name=lowerCamelCase , artifact_url=lowerCamelCase , output_dir=lowerCamelCase , token=lowerCamelCase ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): get_last_daily_ci_artifacts(lowerCamelCase , lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = {} for artifact_name in artifact_names: UpperCAmelCase__ = os.path.join(lowerCamelCase , f'''{artifact_name}.zip''' ) if os.path.isfile(lowerCamelCase ): UpperCAmelCase__ = {} with zipfile.ZipFile(lowerCamelCase ) as z: for filename in z.namelist(): if not os.path.isdir(lowerCamelCase ): # read the file with z.open(lowerCamelCase ) as f: UpperCAmelCase__ = f.read().decode('UTF-8' ) return results
711
"""simple docstring""" import random class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : str ): UpperCAmelCase__ = [ord(lowerCamelCase__ ) for i in text] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i in plain: UpperCAmelCase__ = random.randint(1 ,300 ) UpperCAmelCase__ = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): UpperCAmelCase__ = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ : Dict = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
632
0
"""simple docstring""" import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : int = get_tests_dir('fixtures/spiece.model') @require_sentencepiece @require_tokenizers class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = DebertaVaTokenizer snake_case__ = DebertaVaTokenizerFast snake_case__ = True snake_case__ = True def __lowerCAmelCase ( self : Tuple ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,unk_token='<unk>' ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : Tuple ): UpperCAmelCase__ = 'this is a test' UpperCAmelCase__ = 'this is a test' return input_text, output_text def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = '<pad>' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'<pad>' ) self.assertEqual(vocab_keys[1] ,'<unk>' ) self.assertEqual(vocab_keys[-1] ,'[PAD]' ) self.assertEqual(len(lowerCamelCase__ ) ,30_001 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,30_000 ) def __lowerCAmelCase ( self : Dict ): # fmt: off UpperCAmelCase__ = ' \tHeLLo!how \n Are yoU? ' UpperCAmelCase__ = ['▁hello', '!', 'how', '▁are', '▁you', '?'] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.' ) def __lowerCAmelCase ( self : List[str] ): pass @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.' ) def __lowerCAmelCase ( self : int ): pass def __lowerCAmelCase ( self : Optional[Any] ): # fmt: off UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): # fmt: off UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): # fmt: off UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): # fmt: off UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : int ): # fmt: off UpperCAmelCase__ = ' \tHeLLo!how \n Are yoU? ' UpperCAmelCase__ = ['▁', '<unk>', 'e', '<unk>', 'o', '!', 'how', '▁', '<unk>', 're', '▁yo', '<unk>', '?'] # fmt: on UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,split_by_punct=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_tokenizer() UpperCAmelCase__ = self.get_rust_tokenizer() UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = self.get_rust_tokenizer() UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = 'This is a test' UpperCAmelCase__ = [13, 1, 4_398, 25, 21, 1_289] UpperCAmelCase__ = ['▁', 'T', 'his', '▁is', '▁a', '▁test'] UpperCAmelCase__ = ['▁', '<unk>', 'his', '▁is', '▁a', '▁test'] UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = DebertaVaTokenizerFast(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) # fmt: off UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = [13, 1, 23, 386, 19, 561, 3_050, 15, 17, 48, 25, 8_256, 18, 1, 9] UpperCAmelCase__ = ['▁', 'I', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.', ] UpperCAmelCase__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on UpperCAmelCase__ = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = rust_tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = DebertaVaTokenizer(lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.encode('sequence builders' ) UpperCAmelCase__ = tokenizer.encode('multi-sequence build' ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] ,lowerCamelCase__ ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] ,lowerCamelCase__ ,) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[1, 39_867, 36, 19_390, 486, 27, 35_052, 81_436, 18, 60_685, 1_225, 7, 35_052, 81_436, 18, 9_367, 16_899, 18, 15_937, 53, 594, 773, 18, 16_287, 30_465, 36, 15_937, 6, 41_139, 38, 36_979, 60_763, 191, 6, 34_132, 99, 6, 50_538, 390, 43_230, 6, 34_132, 2_779, 20_850, 14, 699, 1_072, 1_194, 36, 382, 10_901, 53, 7, 699, 1_072, 2_084, 36, 20_422, 630, 53, 19, 105, 3_049, 1_896, 1_053, 16_899, 1_506, 11, 37_978, 4_243, 7, 1_237, 31_869, 200, 16_566, 654, 6, 35_052, 81_436, 7, 55_630, 13_593, 4, 2], [1, 26, 15_011, 13, 667, 8, 1_053, 18, 23_611, 1_237, 72_356, 12_820, 34, 104_134, 1_209, 35, 13_313, 6_627, 21, 202, 347, 7, 164, 2_399, 11, 46, 4_485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1_232, 2_864, 15_785, 14_951, 105, 5, 8_581, 1_250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/deberta-v2-xlarge' ,revision='ad6e42c1532ddf3a15c39246b63f5559d558b670' ,)
712
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
632
0
"""simple docstring""" import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowerCAmelCase__ : str = logging.getLogger(__name__) def a_ ( lowerCamelCase , lowerCamelCase ): # save results if os.path.exists(lowerCamelCase ): if os.path.exists(os.path.join(lowerCamelCase , 'config.json' ) ) and os.path.isfile( os.path.join(lowerCamelCase , 'config.json' ) ): os.remove(os.path.join(lowerCamelCase , 'config.json' ) ) if os.path.exists(os.path.join(lowerCamelCase , 'pytorch_model.bin' ) ) and os.path.isfile( os.path.join(lowerCamelCase , 'pytorch_model.bin' ) ): os.remove(os.path.join(lowerCamelCase , 'pytorch_model.bin' ) ) else: os.makedirs(lowerCamelCase ) model.save_pretrained(lowerCamelCase ) def a_ ( lowerCamelCase , lowerCamelCase=False ): UpperCAmelCase__ = 2 if unlogit: UpperCAmelCase__ = torch.pow(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = p * torch.log(lowerCamelCase ) UpperCAmelCase__ = 0 return -plogp.sum(dim=-1 ) def a_ ( lowerCamelCase ): logger.info('lv, h >\t' + '\t'.join(f'''{x + 1}''' for x in range(len(lowerCamelCase ) ) ) ) for row in range(len(lowerCamelCase ) ): if tensor.dtype != torch.long: logger.info(f'''layer {row + 1}:\t''' + '\t'.join(f'''{x:.5f}''' for x in tensor[row].cpu().data ) ) else: logger.info(f'''layer {row + 1}:\t''' + '\t'.join(f'''{x:d}''' for x in tensor[row].cpu().data ) ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=False ): UpperCAmelCase__ , UpperCAmelCase__ = model.config.num_hidden_layers, model.config.num_attention_heads UpperCAmelCase__ = torch.zeros(lowerCamelCase , lowerCamelCase ).to(args.device ) UpperCAmelCase__ = torch.zeros(lowerCamelCase , lowerCamelCase ).to(args.device ) if head_mask is None: UpperCAmelCase__ = torch.ones(lowerCamelCase , lowerCamelCase ).to(args.device ) head_mask.requires_grad_(requires_grad=lowerCamelCase ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = 0.0 for step, inputs in enumerate(tqdm(lowerCamelCase , desc='Iteration' , disable=args.local_rank not in [-1, 0] ) ): UpperCAmelCase__ = tuple(t.to(args.device ) for t in inputs ) ((UpperCAmelCase__ ) , ) = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) UpperCAmelCase__ = model(lowerCamelCase , labels=lowerCamelCase , head_mask=lowerCamelCase ) # (loss), lm_logits, presents, (all hidden_states), (attentions) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(lowerCamelCase ): UpperCAmelCase__ = entropy(attn.detach() , lowerCamelCase ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(lowerCamelCase ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: UpperCAmelCase__ = 2 UpperCAmelCase__ = torch.pow(torch.pow(lowerCamelCase , lowerCamelCase ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-2_0 if not args.dont_normalize_global_importance: UpperCAmelCase__ = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('Attention entropies' ) print_ad_tensor(lowerCamelCase ) if compute_importance: logger.info('Head importance scores' ) print_ad_tensor(lowerCamelCase ) logger.info('Head ranked by importance scores' ) UpperCAmelCase__ = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) UpperCAmelCase__ = torch.arange( head_importance.numel() , device=args.device ) UpperCAmelCase__ = head_ranks.view_as(lowerCamelCase ) print_ad_tensor(lowerCamelCase ) return attn_entropy, head_importance, total_loss def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = compute_heads_importance(lowerCamelCase , lowerCamelCase , lowerCamelCase , compute_entropy=lowerCamelCase ) UpperCAmelCase__ = 1 / loss # instead of downsteam score use the LM loss logger.info('Pruning: original score: %f, threshold: %f' , lowerCamelCase , original_score * args.masking_threshold ) UpperCAmelCase__ = torch.ones_like(lowerCamelCase ) UpperCAmelCase__ = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) UpperCAmelCase__ = original_score while current_score >= original_score * args.masking_threshold: UpperCAmelCase__ = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads UpperCAmelCase__ = float('Inf' ) UpperCAmelCase__ = head_importance.view(-1 ).sort()[1] if len(lowerCamelCase ) <= num_to_mask: print('BREAK BY num_to_mask' ) break # mask heads UpperCAmelCase__ = current_heads_to_mask[:num_to_mask] logger.info('Heads to mask: %s' , str(current_heads_to_mask.tolist() ) ) UpperCAmelCase__ = new_head_mask.view(-1 ) UpperCAmelCase__ = 0.0 UpperCAmelCase__ = new_head_mask.view_as(lowerCamelCase ) UpperCAmelCase__ = new_head_mask.clone().detach() print_ad_tensor(lowerCamelCase ) # Compute metric and head importance again UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = compute_heads_importance( lowerCamelCase , lowerCamelCase , lowerCamelCase , compute_entropy=lowerCamelCase , head_mask=lowerCamelCase ) UpperCAmelCase__ = 1 / loss logger.info( 'Masking: current score: %f, remaining heads %d (%.1f percents)' , lowerCamelCase , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 1_0_0 , ) logger.info('Final head mask' ) print_ad_tensor(lowerCamelCase ) np.save(os.path.join(args.output_dir , 'head_mask.npy' ) , head_mask.detach().cpu().numpy() ) return head_mask def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = datetime.now() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = compute_heads_importance( lowerCamelCase , lowerCamelCase , lowerCamelCase , compute_entropy=lowerCamelCase , compute_importance=lowerCamelCase , head_mask=lowerCamelCase ) UpperCAmelCase__ = 1 / loss UpperCAmelCase__ = datetime.now() - before_time UpperCAmelCase__ = sum(p.numel() for p in model.parameters() ) UpperCAmelCase__ = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(lowerCamelCase ) ) } for k, v in heads_to_prune.items(): if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = [ v, ] assert sum(len(lowerCamelCase ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(lowerCamelCase ) UpperCAmelCase__ = sum(p.numel() for p in model.parameters() ) UpperCAmelCase__ = datetime.now() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = compute_heads_importance( lowerCamelCase , lowerCamelCase , lowerCamelCase , compute_entropy=lowerCamelCase , compute_importance=lowerCamelCase , head_mask=lowerCamelCase , actually_pruned=lowerCamelCase , ) UpperCAmelCase__ = 1 / loss UpperCAmelCase__ = datetime.now() - before_time logger.info( 'Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)' , lowerCamelCase , lowerCamelCase , pruned_num_params / original_num_params * 1_0_0 , ) logger.info('Pruning: score with masking: %f score with pruning: %f' , lowerCamelCase , lowerCamelCase ) logger.info('Pruning: speed ratio (original timing / new timing): %f percents' , original_time / new_time * 1_0_0 ) save_model(lowerCamelCase , args.output_dir ) def a_ ( ): UpperCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--data_dir' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='The input data dir. Should contain the .tsv files (or other data files) for the task.' , ) parser.add_argument( '--model_name_or_path' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--output_dir' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='The output directory where the model predictions and checkpoints will be written.' , ) # Other parameters parser.add_argument( '--config_name' , default='' , type=lowerCamelCase , help='Pretrained config name or path if not the same as model_name_or_path' , ) parser.add_argument( '--tokenizer_name' , default='' , type=lowerCamelCase , help='Pretrained tokenizer name or path if not the same as model_name_or_path' , ) parser.add_argument( '--cache_dir' , default=lowerCamelCase , type=lowerCamelCase , help='Where do you want to store the pre-trained models downloaded from s3' , ) parser.add_argument( '--data_subset' , type=lowerCamelCase , default=-1 , help='If > 0: limit the data to a subset of data_subset instances.' ) parser.add_argument( '--overwrite_output_dir' , action='store_true' , help='Whether to overwrite data in output directory' ) parser.add_argument( '--overwrite_cache' , action='store_true' , help='Overwrite the cached training and evaluation sets' ) parser.add_argument( '--dont_normalize_importance_by_layer' , action='store_true' , help='Don\'t normalize importance score by layers' ) parser.add_argument( '--dont_normalize_global_importance' , action='store_true' , help='Don\'t normalize all importance scores between 0 and 1' , ) parser.add_argument( '--try_masking' , action='store_true' , help='Whether to try to mask head until a threshold of accuracy.' ) parser.add_argument( '--masking_threshold' , default=0.9 , type=lowerCamelCase , help='masking threshold in term of metrics (stop masking when metric < threshold * original metric value).' , ) parser.add_argument( '--masking_amount' , default=0.1 , type=lowerCamelCase , help='Amount to heads to masking at each masking step.' ) parser.add_argument('--metric_name' , default='acc' , type=lowerCamelCase , help='Metric to use for head masking.' ) parser.add_argument( '--max_seq_length' , default=1_2_8 , type=lowerCamelCase , help=( 'The maximum total input sequence length after WordPiece tokenization. \n' 'Sequences longer than this will be truncated, sequences shorter padded.' ) , ) parser.add_argument('--batch_size' , default=1 , type=lowerCamelCase , help='Batch size.' ) parser.add_argument('--seed' , type=lowerCamelCase , default=4_2 ) parser.add_argument('--local_rank' , type=lowerCamelCase , default=-1 , help='local_rank for distributed training on gpus' ) parser.add_argument('--no_cuda' , action='store_true' , help='Whether not to use CUDA when available' ) parser.add_argument('--server_ip' , type=lowerCamelCase , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=lowerCamelCase , default='' , help='Can be used for distant debugging.' ) UpperCAmelCase__ = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=lowerCamelCase ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: UpperCAmelCase__ = torch.device('cuda' if torch.cuda.is_available() and not args.no_cuda else 'cpu' ) UpperCAmelCase__ = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) UpperCAmelCase__ = torch.device('cuda' , args.local_rank ) UpperCAmelCase__ = 1 torch.distributed.init_process_group(backend='nccl' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('device: {} n_gpu: {}, distributed: {}'.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) UpperCAmelCase__ = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: UpperCAmelCase__ = nn.parallel.DistributedDataParallel( lowerCamelCase , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=lowerCamelCase ) elif args.n_gpu > 1: UpperCAmelCase__ = nn.DataParallel(lowerCamelCase ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=lowerCamelCase ) torch.save(lowerCamelCase , os.path.join(args.output_dir , 'run_args.bin' ) ) logger.info('Training/evaluation parameters %s' , lowerCamelCase ) # Prepare dataset UpperCAmelCase__ = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) UpperCAmelCase__ = (torch.from_numpy(lowerCamelCase ),) UpperCAmelCase__ = TensorDataset(*lowerCamelCase ) UpperCAmelCase__ = RandomSampler(lowerCamelCase ) UpperCAmelCase__ = DataLoader(lowerCamelCase , sampler=lowerCamelCase , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: UpperCAmelCase__ = mask_heads(lowerCamelCase , lowerCamelCase , lowerCamelCase ) prune_heads(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if __name__ == "__main__": main()
713
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Dict ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_torch def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,torch.tensor(lowerCamelCase__ ) ,torch.tensor(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) @require_vision @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_tf def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,tf.convert_to_tensor(lowerCamelCase__ ) ,tf.convert_to_tensor(lowerCamelCase__ ) ,return_tensors='tf' ,) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowerCamelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = np.random.randint(0 ,2 ,size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase__ = [tf.convert_to_tensor(lowerCamelCase__ )] UpperCAmelCase__ = [torch.tensor(lowerCamelCase__ )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='pt' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) )
632
0
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCAmelCase__ : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCAmelCase__ : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowerCAmelCase__ : set[int] = {ord(char) for char in VALID_CHARS} lowerCAmelCase__ : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = '' UpperCAmelCase__ = 4_2 UpperCAmelCase__ = 4_2 UpperCAmelCase__ = 4_2 for keychar, cipherchar in zip(cycle(lowerCamelCase ) , lowerCamelCase ): UpperCAmelCase__ = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCamelCase ) return decoded def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] for key in product(lowerCamelCase , repeat=3 ): UpperCAmelCase__ = try_key(lowerCamelCase , lowerCamelCase ) if encoded is not None: possibles.append(lowerCamelCase ) return possibles def a_ ( lowerCamelCase , lowerCamelCase ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCamelCase = "p059_cipher.txt" ): UpperCAmelCase__ = 4_2 UpperCAmelCase__ = 4_2 UpperCAmelCase__ = 4_2 UpperCAmelCase__ = 4_2 UpperCAmelCase__ = Path(lowerCamelCase ).parent.joinpath(lowerCamelCase ).read_text(encoding='utf-8' ) UpperCAmelCase__ = [int(lowerCamelCase ) for number in data.strip().split(',' )] UpperCAmelCase__ = filter_valid_chars(lowerCamelCase ) for common_word in COMMON_WORDS: UpperCAmelCase__ = filter_common_word(lowerCamelCase , lowerCamelCase ) if len(lowerCamelCase ) == 1: break UpperCAmelCase__ = possibles[0] return sum(ord(lowerCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
714
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Any = logging.get_logger(__name__) lowerCAmelCase__ : str = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "ctrl" snake_case__ = ["past_key_values"] snake_case__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any ,lowerCamelCase__ : str=246_534 ,lowerCamelCase__ : List[str]=256 ,lowerCamelCase__ : Optional[int]=1_280 ,lowerCamelCase__ : Any=8_192 ,lowerCamelCase__ : int=48 ,lowerCamelCase__ : Optional[Any]=16 ,lowerCamelCase__ : Union[str, Any]=0.1 ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : List[str]=1e-6 ,lowerCamelCase__ : List[str]=0.0_2 ,lowerCamelCase__ : Tuple=True ,**lowerCamelCase__ : Optional[Any] ,): UpperCAmelCase__ = vocab_size UpperCAmelCase__ = n_positions UpperCAmelCase__ = n_embd UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = dff UpperCAmelCase__ = resid_pdrop UpperCAmelCase__ = embd_pdrop UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache super().__init__(**lowerCamelCase__ )
632
0
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging lowerCAmelCase__ : List[str] = logging.get_logger(__name__) lowerCAmelCase__ : Tuple = { 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/resolve/main/config.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/config.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/config.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json', } class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "bloom" snake_case__ = ["past_key_values"] snake_case__ = { "num_hidden_layers": "n_layer", "num_attention_heads": "n_head", } def __init__( self : List[Any] ,lowerCamelCase__ : Optional[Any]=250_880 ,lowerCamelCase__ : int=64 ,lowerCamelCase__ : List[Any]=2 ,lowerCamelCase__ : int=8 ,lowerCamelCase__ : int=1e-5 ,lowerCamelCase__ : Dict=0.0_2 ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Optional[int]=1 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : str=False ,lowerCamelCase__ : Optional[Any]=0.0 ,lowerCamelCase__ : List[Any]=0.0 ,lowerCamelCase__ : Any=1 ,lowerCamelCase__ : Tuple=False ,**lowerCamelCase__ : str ,): UpperCAmelCase__ = vocab_size # Backward compatibility with n_embed kwarg UpperCAmelCase__ = kwargs.pop('n_embed' ,lowerCamelCase__ ) UpperCAmelCase__ = hidden_size if n_embed is None else n_embed UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache UpperCAmelCase__ = pretraining_tp UpperCAmelCase__ = apply_residual_connection_post_layernorm UpperCAmelCase__ = hidden_dropout UpperCAmelCase__ = attention_dropout UpperCAmelCase__ = bos_token_id UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = slow_but_exact super().__init__(bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = version.parse("1.12" ) def __init__( self : str ,lowerCamelCase__ : PretrainedConfig ,lowerCamelCase__ : str = "default" ,lowerCamelCase__ : List[PatchingSpec] = None ,lowerCamelCase__ : bool = False ,): super().__init__(lowerCamelCase__ ,task=lowerCamelCase__ ,patching_specs=lowerCamelCase__ ,use_past=lowerCamelCase__ ) if not getattr(self._config ,'pad_token_id' ,lowerCamelCase__ ): # TODO: how to do that better? UpperCAmelCase__ = 0 @property def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(lowerCamelCase__ ,direction='inputs' ,inverted_values_shape=lowerCamelCase__ ) UpperCAmelCase__ = {0: 'batch', 1: 'past_sequence + sequence'} else: UpperCAmelCase__ = {0: 'batch', 1: 'sequence'} return common_inputs @property def __lowerCAmelCase ( self : Dict ): return self._config.n_layer @property def __lowerCAmelCase ( self : Any ): return self._config.n_head @property def __lowerCAmelCase ( self : Any ): return 1e-3 def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : "PreTrainedTokenizer" ,lowerCamelCase__ : int = -1 ,lowerCamelCase__ : int = -1 ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional["TensorType"] = None ,): UpperCAmelCase__ = super(lowerCamelCase__ ,self ).generate_dummy_inputs( lowerCamelCase__ ,batch_size=lowerCamelCase__ ,seq_length=lowerCamelCase__ ,is_pair=lowerCamelCase__ ,framework=lowerCamelCase__ ) # We need to order the input in the way they appears in the forward() UpperCAmelCase__ = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch UpperCAmelCase__ , UpperCAmelCase__ = common_inputs['input_ids'].shape # Not using the same length for past_key_values UpperCAmelCase__ = seqlen + 2 UpperCAmelCase__ = self._config.hidden_size // self.num_attention_heads UpperCAmelCase__ = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) UpperCAmelCase__ = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) UpperCAmelCase__ = [ (torch.zeros(lowerCamelCase__ ), torch.zeros(lowerCamelCase__ )) for _ in range(self.num_layers ) ] UpperCAmelCase__ = common_inputs['attention_mask'] if self.use_past: UpperCAmelCase__ = ordered_inputs['attention_mask'].dtype UpperCAmelCase__ = torch.cat( [ordered_inputs['attention_mask'], torch.ones(lowerCamelCase__ ,lowerCamelCase__ ,dtype=lowerCamelCase__ )] ,dim=1 ) return ordered_inputs @property def __lowerCAmelCase ( self : str ): return 13
715
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCAmelCase__ : Dict = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def a_ ( ): UpperCAmelCase__ = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCAmelCase__ = get_sagemaker_input() else: UpperCAmelCase__ = get_cluster_input() return config def a_ ( lowerCamelCase=None ): if subparsers is not None: UpperCAmelCase__ = subparsers.add_parser('config' , description=lowerCamelCase ) else: UpperCAmelCase__ = argparse.ArgumentParser('Accelerate config command' , description=lowerCamelCase ) parser.add_argument( '--config_file' , default=lowerCamelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def a_ ( lowerCamelCase ): UpperCAmelCase__ = get_user_input() if args.config_file is not None: UpperCAmelCase__ = args.config_file else: if not os.path.isdir(lowerCamelCase ): os.makedirs(lowerCamelCase ) UpperCAmelCase__ = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(lowerCamelCase ) else: config.to_yaml_file(lowerCamelCase ) print(f'''accelerate configuration saved at {config_file}''' ) def a_ ( ): UpperCAmelCase__ = config_command_parser() UpperCAmelCase__ = parser.parse_args() config_command(lowerCamelCase ) if __name__ == "__main__": main()
632
0
"""simple docstring""" def a_ ( lowerCamelCase ): if not isinstance(lowerCamelCase , lowerCamelCase ): raise TypeError('only integers accepted as input' ) else: UpperCAmelCase__ = str(abs(lowerCamelCase ) ) UpperCAmelCase__ = [list(lowerCamelCase ) for char in range(len(lowerCamelCase ) )] for index in range(len(lowerCamelCase ) ): num_transpositions[index].pop(lowerCamelCase ) return max( int(''.join(list(lowerCamelCase ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
716
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): return x if y == 0 else greatest_common_divisor(lowerCamelCase , x % y ) def a_ ( lowerCamelCase , lowerCamelCase ): return (x * y) // greatest_common_divisor(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase = 2_0 ): UpperCAmelCase__ = 1 for i in range(1 , n + 1 ): UpperCAmelCase__ = lcm(lowerCamelCase , lowerCamelCase ) return g if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" from __future__ import annotations def a_ ( lowerCamelCase ): # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(lowerCamelCase ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(lowerCamelCase ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
717
"""simple docstring""" import warnings from functools import wraps from typing import Callable def a_ ( lowerCamelCase ): @wraps(lowerCamelCase ) def _inner_fn(*lowerCamelCase , **lowerCamelCase ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , lowerCamelCase , ) return fn(*lowerCamelCase , **lowerCamelCase ) return _inner_fn
632
0
"""simple docstring""" def a_ ( lowerCamelCase = 6_0_0_8_5_1_4_7_5_1_4_3 ): try: UpperCAmelCase__ = int(lowerCamelCase ) except (TypeError, ValueError): raise TypeError('Parameter n must be int or castable to int.' ) if n <= 0: raise ValueError('Parameter n must be greater than or equal to one.' ) UpperCAmelCase__ = 1 UpperCAmelCase__ = 2 while i * i <= n: while n % i == 0: UpperCAmelCase__ = i n //= i i += 1 if n > 1: UpperCAmelCase__ = n return int(lowerCamelCase ) if __name__ == "__main__": print(F"""{solution() = }""")
718
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCAmelCase__ : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCAmelCase__ : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowerCAmelCase__ : set[int] = {ord(char) for char in VALID_CHARS} lowerCAmelCase__ : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = "" UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 for keychar, cipherchar in zip(cycle(lowerCamelCase ) , lowerCamelCase ): UpperCAmelCase__ = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCamelCase ) return decoded def a_ ( lowerCamelCase ): UpperCAmelCase__ = [] for key in product(lowerCamelCase , repeat=3 ): UpperCAmelCase__ = try_key(lowerCamelCase , lowerCamelCase ) if encoded is not None: possibles.append(lowerCamelCase ) return possibles def a_ ( lowerCamelCase , lowerCamelCase ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCamelCase = "p059_cipher.txt" ): UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = Path(lowerCamelCase ).parent.joinpath(lowerCamelCase ).read_text(encoding='utf-8' ) UpperCAmelCase__ = [int(lowerCamelCase ) for number in data.strip().split(',' )] UpperCAmelCase__ = filter_valid_chars(lowerCamelCase ) for common_word in COMMON_WORDS: UpperCAmelCase__ = filter_common_word(lowerCamelCase , lowerCamelCase ) if len(lowerCamelCase ) == 1: break UpperCAmelCase__ = possibles[0] return sum(ord(lowerCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
632
0
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
719
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = XLMProphetNetTokenizer snake_case__ = False snake_case__ = True def __lowerCAmelCase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = '[PAD]' UpperCAmelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'[PAD]' ) self.assertEqual(vocab_keys[1] ,'[CLS]' ) self.assertEqual(vocab_keys[-1] ,'j' ) self.assertEqual(len(lowerCamelCase__ ) ,1_012 ) def __lowerCAmelCase ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_012 ) def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = XLMProphetNetTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) UpperCAmelCase__ = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) UpperCAmelCase__ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] ,) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] ,) @cached_property def __lowerCAmelCase ( self : Dict ): return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'Hello World!' UpperCAmelCase__ = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowerCAmelCase ( self : List[str] ): # fmt: off UpperCAmelCase__ = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='microsoft/xprophetnet-large-wiki100-cased' ,revision='1acad1643ddd54a44df6a1b797ada8373685d90e' ,)
632
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : str = logging.get_logger(__name__) def a_ ( lowerCamelCase , lowerCamelCase=False ): UpperCAmelCase__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" UpperCAmelCase__ = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase__ = '' else: UpperCAmelCase__ = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase__ = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase__ = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase__ = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase__ = in_proj_bias[: config.hidden_size] UpperCAmelCase__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase__ = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase__ = in_proj_bias[-config.hidden_size :] def a_ ( lowerCamelCase ): UpperCAmelCase__ = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(lowerCamelCase , lowerCamelCase ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = dct.pop(lowerCamelCase ) UpperCAmelCase__ = val def a_ ( ): UpperCAmelCase__ = 'http://images.cocodataset.org/val2017/000000039769.jpg' UpperCAmelCase__ = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ) return im @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = ViTConfig() UpperCAmelCase__ = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": UpperCAmelCase__ = True UpperCAmelCase__ = int(vit_name[-1_2:-1_0] ) UpperCAmelCase__ = int(vit_name[-9:-6] ) else: UpperCAmelCase__ = 1_0_0_0 UpperCAmelCase__ = 'huggingface/label-files' UpperCAmelCase__ = 'imagenet-1k-id2label.json' UpperCAmelCase__ = json.load(open(hf_hub_download(lowerCamelCase , lowerCamelCase , repo_type='dataset' ) , 'r' ) ) UpperCAmelCase__ = {int(lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase__ = idalabel UpperCAmelCase__ = {v: k for k, v in idalabel.items()} UpperCAmelCase__ = int(vit_name[-6:-4] ) UpperCAmelCase__ = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('tiny' ): UpperCAmelCase__ = 1_9_2 UpperCAmelCase__ = 7_6_8 UpperCAmelCase__ = 1_2 UpperCAmelCase__ = 3 elif vit_name[9:].startswith('small' ): UpperCAmelCase__ = 3_8_4 UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 1_2 UpperCAmelCase__ = 6 else: pass else: if vit_name[4:].startswith('small' ): UpperCAmelCase__ = 7_6_8 UpperCAmelCase__ = 2_3_0_4 UpperCAmelCase__ = 8 UpperCAmelCase__ = 8 elif vit_name[4:].startswith('base' ): pass elif vit_name[4:].startswith('large' ): UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 4_0_9_6 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif vit_name[4:].startswith('huge' ): UpperCAmelCase__ = 1_2_8_0 UpperCAmelCase__ = 5_1_2_0 UpperCAmelCase__ = 3_2 UpperCAmelCase__ = 1_6 # load original model from timm UpperCAmelCase__ = timm.create_model(lowerCamelCase , pretrained=lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase__ = timm_model.state_dict() if base_model: remove_classification_head_(lowerCamelCase ) UpperCAmelCase__ = create_rename_keys(lowerCamelCase , lowerCamelCase ) for src, dest in rename_keys: rename_key(lowerCamelCase , lowerCamelCase , lowerCamelCase ) read_in_q_k_v(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": UpperCAmelCase__ = ViTModel(lowerCamelCase ).eval() else: UpperCAmelCase__ = ViTForImageClassification(lowerCamelCase ).eval() model.load_state_dict(lowerCamelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: UpperCAmelCase__ = DeiTImageProcessor(size=config.image_size ) else: UpperCAmelCase__ = ViTImageProcessor(size=config.image_size ) UpperCAmelCase__ = image_processor(images=prepare_img() , return_tensors='pt' ) UpperCAmelCase__ = encoding['pixel_values'] UpperCAmelCase__ = model(lowerCamelCase ) if base_model: UpperCAmelCase__ = timm_model.forward_features(lowerCamelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(lowerCamelCase , outputs.pooler_output , atol=1e-3 ) else: UpperCAmelCase__ = timm_model(lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCamelCase , outputs.logits , atol=1e-3 ) Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) print(f'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_patch16_224', type=str, help='Name of the ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) lowerCAmelCase__ : List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
720
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def a_ ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def a_ ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] lowerCAmelCase__ : str = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') lowerCAmelCase__ : str = sorted({word.strip().lower() for word in data.splitlines()}) lowerCAmelCase__ : Optional[Any] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": lowerCAmelCase__ : int = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
632
0
"""simple docstring""" import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ , UpperCAmelCase__ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' ,from_pt=lowerCamelCase__ ,dtype=jnp.bfloataa ) UpperCAmelCase__ , UpperCAmelCase__ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,controlnet=lowerCamelCase__ ,from_pt=lowerCamelCase__ ,dtype=jnp.bfloataa ) UpperCAmelCase__ = controlnet_params UpperCAmelCase__ = 'bird' UpperCAmelCase__ = jax.device_count() UpperCAmelCase__ = pipe.prepare_text_inputs([prompts] * num_samples ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) UpperCAmelCase__ = pipe.prepare_image_inputs([canny_image] * num_samples ) UpperCAmelCase__ = jax.random.PRNGKey(0 ) UpperCAmelCase__ = jax.random.split(lowerCamelCase__ ,jax.device_count() ) UpperCAmelCase__ = replicate(lowerCamelCase__ ) UpperCAmelCase__ = shard(lowerCamelCase__ ) UpperCAmelCase__ = shard(lowerCamelCase__ ) UpperCAmelCase__ = pipe( prompt_ids=lowerCamelCase__ ,image=lowerCamelCase__ ,params=lowerCamelCase__ ,prng_seed=lowerCamelCase__ ,num_inference_steps=50 ,jit=lowerCamelCase__ ,).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) UpperCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase__ = images[0, 253:256, 253:256, -1] UpperCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase__ = jnp.array( [0.1_6_7_9_6_9, 0.1_1_6_6_9_9, 0.0_8_1_5_4_3, 0.1_5_4_2_9_7, 0.1_3_2_8_1_2, 0.1_0_8_8_8_7, 0.1_6_9_9_2_2, 0.1_6_9_9_2_2, 0.2_0_5_0_7_8] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ , UpperCAmelCase__ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' ,from_pt=lowerCamelCase__ ,dtype=jnp.bfloataa ) UpperCAmelCase__ , UpperCAmelCase__ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,controlnet=lowerCamelCase__ ,from_pt=lowerCamelCase__ ,dtype=jnp.bfloataa ) UpperCAmelCase__ = controlnet_params UpperCAmelCase__ = 'Chef in the kitchen' UpperCAmelCase__ = jax.device_count() UpperCAmelCase__ = pipe.prepare_text_inputs([prompts] * num_samples ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) UpperCAmelCase__ = pipe.prepare_image_inputs([pose_image] * num_samples ) UpperCAmelCase__ = jax.random.PRNGKey(0 ) UpperCAmelCase__ = jax.random.split(lowerCamelCase__ ,jax.device_count() ) UpperCAmelCase__ = replicate(lowerCamelCase__ ) UpperCAmelCase__ = shard(lowerCamelCase__ ) UpperCAmelCase__ = shard(lowerCamelCase__ ) UpperCAmelCase__ = pipe( prompt_ids=lowerCamelCase__ ,image=lowerCamelCase__ ,params=lowerCamelCase__ ,prng_seed=lowerCamelCase__ ,num_inference_steps=50 ,jit=lowerCamelCase__ ,).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) UpperCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase__ = images[0, 253:256, 253:256, -1] UpperCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase__ = jnp.array( [[0.2_7_1_4_8_4, 0.2_6_1_7_1_9, 0.2_7_5_3_9_1, 0.2_7_7_3_4_4, 0.2_7_9_2_9_7, 0.2_9_1_0_1_6, 0.2_9_4_9_2_2, 0.3_0_2_7_3_4, 0.3_0_2_7_3_4]] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
721
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class snake_case ( ctypes.Structure ): """simple docstring""" snake_case__ = [("size", ctypes.c_int), ("visible", ctypes.c_byte)] def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def a_ ( ): if os.name == "nt": UpperCAmelCase__ = CursorInfo() UpperCAmelCase__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) UpperCAmelCase__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def a_ ( ): try: hide_cursor() yield finally: show_cursor()
632
0
def a_ ( lowerCamelCase , lowerCamelCase ): return int(input_a == input_a == 0 ) def a_ ( ): print('Truth Table of NOR Gate:' ) print('| Input 1 | Input 2 | Output |' ) print(f'''| 0 | 0 | {nor_gate(0 , 0 )} |''' ) print(f'''| 0 | 1 | {nor_gate(0 , 1 )} |''' ) print(f'''| 1 | 0 | {nor_gate(1 , 0 )} |''' ) print(f'''| 1 | 1 | {nor_gate(1 , 1 )} |''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
700
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class snake_case ( __UpperCAmelCase ): """simple docstring""" def __get__( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) UpperCAmelCase__ = '__cached_' + self.fget.__name__ UpperCAmelCase__ = getattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) if cached is None: UpperCAmelCase__ = self.fget(lowerCamelCase__ ) setattr(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) return cached def a_ ( lowerCamelCase ): UpperCAmelCase__ = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def a_ ( lowerCamelCase ): if is_torch_fx_proxy(lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return isinstance(lowerCamelCase , np.ndarray ) def a_ ( lowerCamelCase ): return _is_numpy(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.Tensor ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch return isinstance(lowerCamelCase , torch.device ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_device(lowerCamelCase ) def a_ ( lowerCamelCase ): import torch if isinstance(lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) else: return False return isinstance(lowerCamelCase , torch.dtype ) def a_ ( lowerCamelCase ): return False if not is_torch_available() else _is_torch_dtype(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf return isinstance(lowerCamelCase , tf.Tensor ) def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tensorflow(lowerCamelCase ) def a_ ( lowerCamelCase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowerCamelCase , 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowerCamelCase ) return type(lowerCamelCase ) == tf.Tensor def a_ ( lowerCamelCase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowerCamelCase ) def a_ ( lowerCamelCase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowerCamelCase , jnp.ndarray ) def a_ ( lowerCamelCase ): return False if not is_flax_available() else _is_jax(lowerCamelCase ) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return [to_py_obj(lowerCamelCase ) for o in obj] elif is_tf_tensor(lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ).tolist() elif isinstance(lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(lowerCamelCase ) for k, v in obj.items()} elif isinstance(lowerCamelCase , (list, tuple) ): return np.array(lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): return obj.numpy() elif is_torch_tensor(lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowerCamelCase ): return np.asarray(lowerCamelCase ) else: return obj class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = fields(self ) # Safety and consistency checks if not len(lowerCamelCase__ ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) UpperCAmelCase__ = getattr(self ,class_fields[0].name ) UpperCAmelCase__ = all(getattr(self ,field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase__ ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = first_field.items() UpperCAmelCase__ = True else: try: UpperCAmelCase__ = iter(lowerCamelCase__ ) UpperCAmelCase__ = True except TypeError: UpperCAmelCase__ = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase__ ): if ( not isinstance(lowerCamelCase__ ,(list, tuple) ) or not len(lowerCamelCase__ ) == 2 or not isinstance(element[0] ,lowerCamelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute UpperCAmelCase__ = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self ,element[0] ,element[1] ) if element[1] is not None: UpperCAmelCase__ = element[1] elif first_field is not None: UpperCAmelCase__ = first_field else: for field in class_fields: UpperCAmelCase__ = getattr(self ,field.name ) if v is not None: UpperCAmelCase__ = v def __delitem__( self : List[Any] ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : Dict ): raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : str ,*lowerCamelCase__ : int ,**lowerCamelCase__ : Optional[Any] ): raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : List[Any] ,*lowerCamelCase__ : int ,**lowerCamelCase__ : List[Any] ): raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def __lowerCAmelCase ( self : Optional[int] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Optional[int] ): raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[Any] ): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): UpperCAmelCase__ = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __setitem__( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ): # Will raise a KeyException if needed super().__setitem__(lowerCamelCase__ ,lowerCamelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : List[str] ): return tuple(self[k] for k in self.keys() ) class snake_case ( __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" @classmethod def __lowerCAmelCase ( cls : str ,lowerCamelCase__ : Optional[int] ): raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "longest" snake_case__ = "max_length" snake_case__ = "do_not_pad" class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "pt" snake_case__ = "tf" snake_case__ = "np" snake_case__ = "jax" class snake_case : """simple docstring""" def __init__( self : int ,lowerCamelCase__ : List[ContextManager] ): UpperCAmelCase__ = context_managers UpperCAmelCase__ = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase__ ) def __exit__( self : List[Any] ,*lowerCamelCase__ : List[str] ,**lowerCamelCase__ : Dict ): self.stack.__exit__(*lowerCamelCase__ ,**lowerCamelCase__ ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( lowerCamelCase ): UpperCAmelCase__ = model_class.__name__ UpperCAmelCase__ = infer_framework(lowerCamelCase ) if framework == "tf": UpperCAmelCase__ = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": UpperCAmelCase__ = inspect.signature(model_class.forward ) # PyTorch models else: UpperCAmelCase__ = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( lowerCamelCase , lowerCamelCase = "" , lowerCamelCase = "." ): def _flatten_dict(lowerCamelCase , lowerCamelCase="" , lowerCamelCase="." ): for k, v in d.items(): UpperCAmelCase__ = str(lowerCamelCase ) + delimiter + str(lowerCamelCase ) if parent_key else k if v and isinstance(lowerCamelCase , lowerCamelCase ): yield from flatten_dict(lowerCamelCase , lowerCamelCase , delimiter=lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) @contextmanager def a_ ( lowerCamelCase , lowerCamelCase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.transpose(lowerCamelCase , axes=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.T if axes is None else array.permute(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.transpose(lowerCamelCase , perm=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.transpose(lowerCamelCase , axes=lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.reshape(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.reshape(*lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.reshape(lowerCamelCase , lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.reshape(lowerCamelCase , lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase=None ): if is_numpy_array(lowerCamelCase ): return np.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.squeeze(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.squeeze(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.expand_dims(lowerCamelCase , lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.unsqueeze(dim=lowerCamelCase ) elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.expand_dims(lowerCamelCase , axis=lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return jnp.expand_dims(lowerCamelCase , axis=lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase ): if is_numpy_array(lowerCamelCase ): return np.size(lowerCamelCase ) elif is_torch_tensor(lowerCamelCase ): return array.numel() elif is_tf_tensor(lowerCamelCase ): import tensorflow as tf return tf.size(lowerCamelCase ) elif is_jax_tensor(lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowerCamelCase )}.''' ) def a_ ( lowerCamelCase , lowerCamelCase ): for key, value in auto_map.items(): if isinstance(lowerCamelCase , (tuple, list) ): UpperCAmelCase__ = [f'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: UpperCAmelCase__ = f'''{repo_id}--{value}''' return auto_map def a_ ( lowerCamelCase ): for base_class in inspect.getmro(lowerCamelCase ): UpperCAmelCase__ = base_class.__module__ UpperCAmelCase__ = base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
632
0
"""simple docstring""" def a_ ( lowerCamelCase ): for i in range(0 , lowerCamelCase ): for _ in range(0 , n - i - 1 ): # printing spaces print(' ' , end='' ) for _ in range(0 , i + 1 ): # printing stars print('* ' , end='' ) print() def a_ ( lowerCamelCase ): for i in range(lowerCamelCase , 0 , -1 ): for _ in range(lowerCamelCase , 0 , -1 ): # printing stars print('* ' , end='' ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(' ' , end='' ) def a_ ( lowerCamelCase ): if n <= 0: print(' ... .... nothing printing :(' ) return floyd(lowerCamelCase ) # upper half reverse_floyd(lowerCamelCase ) # lower half if __name__ == "__main__": print(r'| /\ | |- | |- |--| |\ /| |-') print(r'|/ \| |- |_ |_ |__| | \/ | |_') lowerCAmelCase__ : Optional[int] = 1 while K: lowerCAmelCase__ : Optional[Any] = int(input('enter the number and , and see the magic : ')) print() pretty_print(user_number) lowerCAmelCase__ : Optional[int] = int(input('press 0 to exit... and 1 to continue...')) print('Good Bye...')
701
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ : Union[str, Any] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained(lowerCamelCase__ ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = generator.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt='first prompt' ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='numpy' ,).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' ,torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = 'cyberpunk 2077' UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.dual_guided( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,text_to_image_strength=0.7_5 ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = 'A painting of a squirrel eating a burger ' UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe.text_to_image( prompt=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCAmelCase__ = pipe.image_variation(lowerCamelCase__ ,generator=lowerCamelCase__ ,output_type='numpy' ).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
632
0
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = VideoToVideoSDPipeline snake_case__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"video"} ) - {"image", "width", "height"} snake_case__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"video"} ) - {"image"} snake_case__ = PipelineTesterMixin.required_optional_params - {"latents"} snake_case__ = False # No `output_type`. snake_case__ = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def __lowerCAmelCase ( self : List[Any] ): torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') ,up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') ,cross_attention_dim=32 ,attention_head_dim=4 ,) UpperCAmelCase__ = DDIMScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule='scaled_linear' ,clip_sample=lowerCamelCase__ ,set_alpha_to_one=lowerCamelCase__ ,) torch.manual_seed(0 ) UpperCAmelCase__ = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=4 ,sample_size=128 ,) torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,hidden_act='gelu' ,projection_dim=512 ,) UpperCAmelCase__ = CLIPTextModel(lowerCamelCase__ ) UpperCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) UpperCAmelCase__ = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def __lowerCAmelCase ( self : str ,lowerCamelCase__ : Any ,lowerCamelCase__ : Union[str, Any]=0 ): # 3 frames UpperCAmelCase__ = floats_tensor((1, 3, 3, 32, 32) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'prompt': 'A painting of a squirrel eating a burger', 'video': video, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = 'cpu' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = VideoToVideoSDPipeline(**lowerCamelCase__ ) UpperCAmelCase__ = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = self.get_dummy_inputs(lowerCamelCase__ ) UpperCAmelCase__ = 'np' UpperCAmelCase__ = sd_pipe(**lowerCamelCase__ ).frames UpperCAmelCase__ = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase__ = np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() ,reason='XFormers attention is only available with CUDA and `xformers` installed' ,) def __lowerCAmelCase ( self : Union[str, Any] ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase__ ,expected_max_diff=5e-3 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def __lowerCAmelCase ( self : Optional[Any] ): pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def __lowerCAmelCase ( self : List[Any] ): pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def __lowerCAmelCase ( self : int ): pass def __lowerCAmelCase ( self : int ): return super().test_progress_bar() @slow @skip_mps class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = VideoToVideoSDPipeline.from_pretrained('cerspense/zeroscope_v2_XL' ,torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) UpperCAmelCase__ = torch.randn((1, 10, 3, 1_024, 576) ,generator=lowerCamelCase__ ) UpperCAmelCase__ = video.to('cuda' ) UpperCAmelCase__ = 'Spiderman is surfing' UpperCAmelCase__ = pipe(lowerCamelCase__ ,video=lowerCamelCase__ ,generator=lowerCamelCase__ ,num_inference_steps=3 ,output_type='pt' ).frames UpperCAmelCase__ = np.array([-1.0_4_5_8_9_8_4, -1.1_2_7_9_2_9_7, -0.9_6_6_3_0_8_6, -0.9_1_5_0_3_9_0_6, -0.7_5_0_9_7_6_5_6] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
702
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
632
0
"""simple docstring""" import unittest from dataclasses import dataclass import pytest from accelerate.commands.config.config_args import SageMakerConfig from accelerate.utils import ComputeEnvironment from accelerate.utils.launch import _convert_nargs_to_dict @dataclass class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ComputeEnvironment.AMAZON_SAGEMAKER snake_case__ = True snake_case__ = "ml.p3.2xlarge" snake_case__ = "accelerate_sagemaker_execution_role" snake_case__ = "hf-sm" snake_case__ = "us-east-1" snake_case__ = 1 snake_case__ = "accelerate-sagemaker-1" snake_case__ = "1.6" snake_case__ = "4.4" snake_case__ = "train.py" snake_case__ = [ "--model_name_or_path", "bert", "--do_train", "False", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] snake_case__ = [ "--model_name_or_path", "bert", "--do_train", "--do_test", "False", "--do_predict", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[int] ): # If no defaults are changed, `to_kwargs` returns an empty dict. UpperCAmelCase__ = _convert_nargs_to_dict(MockLaunchConfig.success_training_script_args ) assert isinstance(converted_args['model_name_or_path'] ,lowerCamelCase__ ) assert isinstance(converted_args['do_train'] ,lowerCamelCase__ ) assert isinstance(converted_args['epochs'] ,lowerCamelCase__ ) assert isinstance(converted_args['learning_rate'] ,lowerCamelCase__ ) assert isinstance(converted_args['max_steps'] ,lowerCamelCase__ ) with pytest.raises(lowerCamelCase__ ): _convert_nargs_to_dict(MockLaunchConfig.fail_training_script_args )
703
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowerCAmelCase__ : str = 'base_with_context' def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['token_embedder']['embedding'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['input_proj']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = ly_weight['attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['encoder_norm']['scale'] ) ) return model def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['time_emb_dense1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['Embed_0']['embedding'] ) , requires_grad=lowerCamelCase ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights['continuous_inputs_projection']['kernel'].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f'''layers_{lyr_num}'''] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_self_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_0']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['self_attention'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = ly_weight['MultiHeadDotProductAttention_0'] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['query']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['key']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['value']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights['out']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['pre_cross_attention_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['pre_mlp_layer_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight['FiLMLayer_1']['DenseGeneral_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_0']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wi_1']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight['mlp']['wo']['kernel'].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['decoder_norm']['scale'] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights['spec_out_dense']['kernel'].T ) ) return model def a_ ( lowerCamelCase ): UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array , lowerCamelCase ) UpperCAmelCase__ = [ 'from __gin__ import dynamic_registration', 'from music_spectrogram_diffusion.models.diffusion import diffusion_utils', 'diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0', 'diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()', ] UpperCAmelCase__ = os.path.join(args.checkpoint_path , '..' , 'config.gin' ) UpperCAmelCase__ = inference.parse_training_gin_file(lowerCamelCase , lowerCamelCase ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path , lowerCamelCase ) UpperCAmelCase__ = DDPMScheduler(beta_schedule='squaredcos_cap_v2' , variance_type='fixed_large' ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['inputs'] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['targets_context'] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='gated-gelu' , ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['targets_context'] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint['target']['token_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint['target']['continuous_encoder'] , lowerCamelCase ) UpperCAmelCase__ = load_decoder(ta_checkpoint['target']['decoder'] , lowerCamelCase ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained('kashif/soundstream_mel_decoder' ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=lowerCamelCase , continuous_encoder=lowerCamelCase , decoder=lowerCamelCase , scheduler=lowerCamelCase , melgan=lowerCamelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=F"""{MODEL}/checkpoint_500000""", type=str, required=False, help='Path to the original jax model checkpoint.', ) lowerCAmelCase__ : List[str] = parser.parse_args() main(args)
632
0
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" snake_case__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : str ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : int = 50_257 ,lowerCamelCase__ : int = 1_024 ,lowerCamelCase__ : int = 768 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : int = 12 ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : str = "gelu_new" ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 0.1 ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : float = 0.0_2 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : bool = False ,): super().__init__() UpperCAmelCase__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) UpperCAmelCase__ = prefix_inner_dim UpperCAmelCase__ = prefix_hidden_dim UpperCAmelCase__ = ( nn.Linear(self.prefix_inner_dim ,self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = ( nn.Linear(self.prefix_hidden_dim ,lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) UpperCAmelCase__ = GPTaConfig( vocab_size=lowerCamelCase__ ,n_positions=lowerCamelCase__ ,n_embd=lowerCamelCase__ ,n_layer=lowerCamelCase__ ,n_head=lowerCamelCase__ ,n_inner=lowerCamelCase__ ,activation_function=lowerCamelCase__ ,resid_pdrop=lowerCamelCase__ ,embd_pdrop=lowerCamelCase__ ,attn_pdrop=lowerCamelCase__ ,layer_norm_epsilon=lowerCamelCase__ ,initializer_range=lowerCamelCase__ ,scale_attn_weights=lowerCamelCase__ ,use_cache=lowerCamelCase__ ,scale_attn_by_inverse_layer_idx=lowerCamelCase__ ,reorder_and_upcast_attn=lowerCamelCase__ ,) UpperCAmelCase__ = GPTaLMHeadModel(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : torch.Tensor ,lowerCamelCase__ : Optional[torch.Tensor] = None ,lowerCamelCase__ : Optional[torch.Tensor] = None ,): UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) UpperCAmelCase__ = self.encode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = self.decode_prefix(lowerCamelCase__ ) UpperCAmelCase__ = torch.cat((prefix_embeds, embedding_text) ,dim=1 ) if labels is not None: UpperCAmelCase__ = self.get_dummy_token(input_ids.shape[0] ,input_ids.device ) UpperCAmelCase__ = torch.cat((dummy_token, input_ids) ,dim=1 ) UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ,labels=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : torch.device ): return torch.zeros(lowerCamelCase__ ,self.prefix_length ,dtype=torch.intaa ,device=lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : List[str] ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[int] ): UpperCAmelCase__ = torch.split(lowerCamelCase__ ,1 ,dim=0 ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for feature in features: UpperCAmelCase__ = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now UpperCAmelCase__ , UpperCAmelCase__ = self.generate_beam( input_embeds=lowerCamelCase__ ,device=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) UpperCAmelCase__ = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 67 ,lowerCamelCase__ : float = 1.0 ,lowerCamelCase__ : Optional[int] = None ,): UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = torch.ones(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.int ) UpperCAmelCase__ = torch.zeros(lowerCamelCase__ ,device=lowerCamelCase__ ,dtype=torch.bool ) if input_embeds is not None: UpperCAmelCase__ = input_embeds else: UpperCAmelCase__ = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): UpperCAmelCase__ = self.transformer(inputs_embeds=lowerCamelCase__ ) UpperCAmelCase__ = outputs.logits UpperCAmelCase__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) UpperCAmelCase__ = logits.softmax(-1 ).log() if scores is None: UpperCAmelCase__ , UpperCAmelCase__ = logits.topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = generated.expand(lowerCamelCase__ ,*generated.shape[1:] ) UpperCAmelCase__ , UpperCAmelCase__ = next_tokens.permute(1 ,0 ), scores.squeeze(0 ) if tokens is None: UpperCAmelCase__ = next_tokens else: UpperCAmelCase__ = tokens.expand(lowerCamelCase__ ,*tokens.shape[1:] ) UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) else: UpperCAmelCase__ = -float(np.inf ) UpperCAmelCase__ = 0 UpperCAmelCase__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 UpperCAmelCase__ = scores_sum / seq_lengths[:, None] UpperCAmelCase__ , UpperCAmelCase__ = scores_sum_average.view(-1 ).topk(lowerCamelCase__ ,-1 ) UpperCAmelCase__ = next_tokens // scores_sum.shape[1] UpperCAmelCase__ = seq_lengths[next_tokens_source] UpperCAmelCase__ = next_tokens % scores_sum.shape[1] UpperCAmelCase__ = next_tokens.unsqueeze(1 ) UpperCAmelCase__ = tokens[next_tokens_source] UpperCAmelCase__ = torch.cat((tokens, next_tokens) ,dim=1 ) UpperCAmelCase__ = generated[next_tokens_source] UpperCAmelCase__ = scores_sum_average * seq_lengths UpperCAmelCase__ = is_stopped[next_tokens_source] UpperCAmelCase__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] ,1 ,-1 ) UpperCAmelCase__ = torch.cat((generated, next_token_embed) ,dim=1 ) UpperCAmelCase__ = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break UpperCAmelCase__ = scores / seq_lengths UpperCAmelCase__ = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length UpperCAmelCase__ = [tokens[i] for i in order] UpperCAmelCase__ = torch.stack(lowerCamelCase__ ,dim=0 ) UpperCAmelCase__ = torch.tensor([seq_lengths[i] for i in order] ,dtype=seq_lengths.dtype ) return output_texts, seq_lengths
704
"""simple docstring""" import socket def a_ ( ): UpperCAmelCase__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase__ = socket.gethostname() UpperCAmelCase__ = 1_2_3_1_2 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: UpperCAmelCase__ = sock.recv(1_0_2_4 ) if not data: break out_file.write(lowerCamelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
632
0
"""simple docstring""" import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. lowerCAmelCase__ : List[str] = abspath(join(dirname(dirname(dirname(__file__))), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def a_ ( lowerCamelCase ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowerCamelCase ) def a_ ( lowerCamelCase ): from transformers.testing_utils import pytest_terminal_summary_main UpperCAmelCase__ = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(lowerCamelCase , id=lowerCamelCase )
705
"""simple docstring""" from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Dict ,lowerCamelCase__ : list[list[int]] ): UpperCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(lowerCamelCase__ ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(lowerCamelCase__ ) != cols: raise error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def __lowerCAmelCase ( self : Union[str, Any] ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __lowerCAmelCase ( self : str ): return len(self.rows ) @property def __lowerCAmelCase ( self : List[Any] ): return len(self.rows[0] ) @property def __lowerCAmelCase ( self : Any ): return (self.num_rows, self.num_columns) @property def __lowerCAmelCase ( self : Optional[int] ): return self.order[0] == self.order[1] def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : str ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __lowerCAmelCase ( self : List[str] ): return bool(self.determinant() ) def __lowerCAmelCase ( self : Any ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(lowerCamelCase__ ).determinant() def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): if (row + column) % 2 == 0: return self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) return -1 * self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): return Matrix( [ [self.get_minor(lowerCamelCase__ ,lowerCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __lowerCAmelCase ( self : int ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : List[str] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(lowerCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in row: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(lowerCamelCase__ ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : int | None = None ): UpperCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise type_error for value in column: if not isinstance(lowerCamelCase__ ,(int, float) ): raise type_error if len(lowerCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : List[Any] ,lowerCamelCase__ : object ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : Any ,lowerCamelCase__ : object ): return not self == other def __neg__( self : Dict ): return self * -1 def __add__( self : str ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : List[str] ,lowerCamelCase__ : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] ,lowerCamelCase__ : Matrix | int | float ): if isinstance(lowerCamelCase__ ,(int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(lowerCamelCase__ ,lowerCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Optional[int] ,lowerCamelCase__ : int ): if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): return sum(row[i] * column[i] for i in range(len(lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
632
0
"""simple docstring""" import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = KandinskyVaaPriorPipeline snake_case__ = ["prompt"] snake_case__ = ["prompt", "negative_prompt"] snake_case__ = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] snake_case__ = False @property def __lowerCAmelCase ( self : List[str] ): return 32 @property def __lowerCAmelCase ( self : str ): return 32 @property def __lowerCAmelCase ( self : List[Any] ): return self.time_input_dim @property def __lowerCAmelCase ( self : Dict ): return self.time_input_dim * 4 @property def __lowerCAmelCase ( self : List[Any] ): return 100 @property def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __lowerCAmelCase ( self : Optional[int] ): torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=self.text_embedder_hidden_size ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,) return CLIPTextModelWithProjection(lowerCamelCase__ ) @property def __lowerCAmelCase ( self : int ): torch.manual_seed(0 ) UpperCAmelCase__ = { 'num_attention_heads': 2, 'attention_head_dim': 12, 'embedding_dim': self.text_embedder_hidden_size, 'num_layers': 1, } UpperCAmelCase__ = PriorTransformer(**lowerCamelCase__ ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 UpperCAmelCase__ = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def __lowerCAmelCase ( self : Optional[int] ): torch.manual_seed(0 ) UpperCAmelCase__ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size ,image_size=224 ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=37 ,num_attention_heads=4 ,num_channels=3 ,num_hidden_layers=5 ,patch_size=14 ,) UpperCAmelCase__ = CLIPVisionModelWithProjection(lowerCamelCase__ ) return model @property def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = CLIPImageProcessor( crop_size=224 ,do_center_crop=lowerCamelCase__ ,do_normalize=lowerCamelCase__ ,do_resize=lowerCamelCase__ ,image_mean=[0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] ,image_std=[0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] ,resample=3 ,size=224 ,) return image_processor def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.dummy_prior UpperCAmelCase__ = self.dummy_image_encoder UpperCAmelCase__ = self.dummy_text_encoder UpperCAmelCase__ = self.dummy_tokenizer UpperCAmelCase__ = self.dummy_image_processor UpperCAmelCase__ = UnCLIPScheduler( variance_type='fixed_small_log' ,prediction_type='sample' ,num_train_timesteps=1_000 ,clip_sample=lowerCamelCase__ ,clip_sample_range=10.0 ,) UpperCAmelCase__ = { 'prior': prior, 'image_encoder': image_encoder, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'scheduler': scheduler, 'image_processor': image_processor, } return components def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Any=0 ): if str(lowerCamelCase__ ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(lowerCamelCase__ ) else: UpperCAmelCase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) UpperCAmelCase__ = { 'prompt': 'horse', 'generator': generator, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = 'cpu' UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**lowerCamelCase__ ) UpperCAmelCase__ = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = pipe(**self.get_dummy_inputs(lowerCamelCase__ ) ) UpperCAmelCase__ = output.image_embeds UpperCAmelCase__ = pipe( **self.get_dummy_inputs(lowerCamelCase__ ) ,return_dict=lowerCamelCase__ ,)[0] UpperCAmelCase__ = image[0, -10:] UpperCAmelCase__ = image_from_tuple[0, -10:] assert image.shape == (1, 32) UpperCAmelCase__ = np.array( [-0.0_5_3_2, 1.7_1_2_0, 0.3_6_5_6, -1.0_8_5_2, -0.8_9_4_6, -1.1_7_5_6, 0.4_3_4_8, 0.2_4_8_2, 0.5_1_4_6, -0.1_1_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = torch_device == 'cpu' UpperCAmelCase__ = True UpperCAmelCase__ = False self._test_inference_batch_single_identical( test_max_difference=lowerCamelCase__ ,relax_max_difference=lowerCamelCase__ ,test_mean_pixel_difference=lowerCamelCase__ ,) @skip_mps def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = torch_device == 'cpu' UpperCAmelCase__ = False self._test_attention_slicing_forward_pass( test_max_difference=lowerCamelCase__ ,test_mean_pixel_difference=lowerCamelCase__ ,)
706
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ : int = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Union[str, Any] = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[Any] = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
632
0
"""simple docstring""" import json import sys def a_ ( lowerCamelCase , lowerCamelCase ): with open(lowerCamelCase , encoding='utf-8' ) as f: UpperCAmelCase__ = json.load(lowerCamelCase ) UpperCAmelCase__ = ['<details>', '<summary>Show updated benchmarks!</summary>', ' '] for benchmark_name in sorted(lowerCamelCase ): UpperCAmelCase__ = results[benchmark_name] UpperCAmelCase__ = benchmark_name.split('/' )[-1] output_md.append(f'''### Benchmark: {benchmark_file_name}''' ) UpperCAmelCase__ = '| metric |' UpperCAmelCase__ = '|--------|' UpperCAmelCase__ = '| new / old (diff) |' for metric_name in sorted(lowerCamelCase ): UpperCAmelCase__ = benchmark_res[metric_name] UpperCAmelCase__ = metric_vals['new'] UpperCAmelCase__ = metric_vals.get('old' , lowerCamelCase ) UpperCAmelCase__ = metric_vals.get('diff' , lowerCamelCase ) UpperCAmelCase__ = f''' {new_val:f}''' if isinstance(lowerCamelCase , (int, float) ) else 'None' if old_val is not None: val_str += f''' / {old_val:f}''' if isinstance(lowerCamelCase , (int, float) ) else "None" if dif_val is not None: val_str += f''' ({dif_val:f})''' if isinstance(lowerCamelCase , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('</details>' ) with open(lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.writelines('\n'.join(lowerCamelCase ) ) if __name__ == "__main__": lowerCAmelCase__ : Optional[int] = sys.argv[1] lowerCAmelCase__ : Dict = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
707
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase__ : Optional[int] = False class snake_case ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) UpperCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( image=lowerCamelCase__ ,generator=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=50 ,output_type='numpy' ,).images UpperCAmelCase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
632
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowerCAmelCase__ : Any = logging.get_logger(__name__) class snake_case ( __UpperCAmelCase ): """simple docstring""" def __init__( self : Any ,*lowerCamelCase__ : List[Any] ,**lowerCamelCase__ : Union[str, Any] ): warnings.warn( 'The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PoolFormerImageProcessor instead.' ,lowerCamelCase__ ,) super().__init__(*lowerCamelCase__ ,**lowerCamelCase__ )
708
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def a_ ( lowerCamelCase = 2_0_0_0_0_0_0 ): UpperCAmelCase__ = [0] UpperCAmelCase__ = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target UpperCAmelCase__ = 0 # the area corresponding to the grid that gives the product closest to target UpperCAmelCase__ = 0 # an estimate of b, using the quadratic formula UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the largest integer less than b_estimate UpperCAmelCase__ = 42 # the triangle number corresponding to b_floor UpperCAmelCase__ = 42 # the triangle number corresponding to b_ceil UpperCAmelCase__ = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): UpperCAmelCase__ = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 UpperCAmelCase__ = floor(lowerCamelCase ) UpperCAmelCase__ = ceil(lowerCamelCase ) UpperCAmelCase__ = triangle_numbers[b_floor] UpperCAmelCase__ = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_first_guess * triangle_a UpperCAmelCase__ = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): UpperCAmelCase__ = triangle_b_second_guess * triangle_a UpperCAmelCase__ = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
632
0
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class snake_case : """simple docstring""" def __init__( self : str ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Union[str, Any]=13 ,lowerCamelCase__ : Union[str, Any]=7 ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : str=True ,lowerCamelCase__ : Optional[int]=True ,lowerCamelCase__ : List[str]=99 ,lowerCamelCase__ : Optional[int]=32 ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[Any]=4 ,lowerCamelCase__ : Optional[Any]=37 ,lowerCamelCase__ : Optional[Any]="gelu" ,lowerCamelCase__ : Optional[int]=0.1 ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : Dict=512 ,lowerCamelCase__ : Optional[int]=16 ,lowerCamelCase__ : Optional[Any]=2 ,lowerCamelCase__ : Optional[int]=0.0_2 ,lowerCamelCase__ : str=False ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : int="None" ,lowerCamelCase__ : Dict=3 ,lowerCamelCase__ : List[Any]=4 ,lowerCamelCase__ : List[str]=None ,): UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_input_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_labels UpperCAmelCase__ = num_choices UpperCAmelCase__ = relative_attention UpperCAmelCase__ = position_biased_input UpperCAmelCase__ = pos_att_type UpperCAmelCase__ = scope def __lowerCAmelCase ( self : str ): UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ = None if self.use_input_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase__ = DebertaVaConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,relative_attention=self.relative_attention ,position_biased_input=self.position_biased_input ,initializer_range=self.initializer_range ,return_dict=lowerCamelCase__ ,) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = TFDebertaVaModel(config=lowerCamelCase__ ) UpperCAmelCase__ = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} UpperCAmelCase__ = [input_ids, input_mask] UpperCAmelCase__ = model(lowerCamelCase__ ) UpperCAmelCase__ = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Tuple ): UpperCAmelCase__ = TFDebertaVaForMaskedLM(config=lowerCamelCase__ ) UpperCAmelCase__ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase__ = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Union[str, Any] ): UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFDebertaVaForSequenceClassification(config=lowerCamelCase__ ) UpperCAmelCase__ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase__ = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : str ): UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFDebertaVaForTokenClassification(config=lowerCamelCase__ ) UpperCAmelCase__ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase__ = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Dict ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Any ,lowerCamelCase__ : Any ): UpperCAmelCase__ = TFDebertaVaForQuestionAnswering(config=lowerCamelCase__ ) UpperCAmelCase__ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase__ = model(lowerCamelCase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = config_and_inputs UpperCAmelCase__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) snake_case__ = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) snake_case__ = False snake_case__ = False def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = TFDebertaVaModelTester(self ) UpperCAmelCase__ = ConfigTester(self ,config_class=lowerCamelCase__ ,hidden_size=37 ) def __lowerCAmelCase ( self : Dict ): self.config_tester.run_common_tests() def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase__ ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase__ ) @slow def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge' ) self.assertIsNotNone(lowerCamelCase__ ) @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason='Model not available yet' ) def __lowerCAmelCase ( self : Any ): pass @slow def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge' ) UpperCAmelCase__ = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) UpperCAmelCase__ = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] UpperCAmelCase__ = tf.constant( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1e-4 )
709
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : Optional[Any] = ['model.decoder.embed_positions.weights'] def a_ ( lowerCamelCase ): if "emb" in name: UpperCAmelCase__ = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: UpperCAmelCase__ = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: UpperCAmelCase__ = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: UpperCAmelCase__ = name.replace('linear1' , 'fc1' ) if "linear2" in name: UpperCAmelCase__ = name.replace('linear2' , 'fc2' ) if "norm1" in name: UpperCAmelCase__ = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: UpperCAmelCase__ = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: UpperCAmelCase__ = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: UpperCAmelCase__ = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: UpperCAmelCase__ = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase__ = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = list(state_dict.keys() ) UpperCAmelCase__ = {} for key in keys: UpperCAmelCase__ = state_dict.pop(lowerCamelCase ) UpperCAmelCase__ = rename_keys(lowerCamelCase ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase__ = val[:hidden_size, :] UpperCAmelCase__ = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase__ = val else: UpperCAmelCase__ = val return state_dict, enc_dec_proj_state_dict def a_ ( lowerCamelCase ): if checkpoint == "small": # default config values UpperCAmelCase__ = 1_0_2_4 UpperCAmelCase__ = 2_4 UpperCAmelCase__ = 1_6 elif checkpoint == "medium": UpperCAmelCase__ = 1_5_3_6 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 2_4 elif checkpoint == "large": UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 4_8 UpperCAmelCase__ = 3_2 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) UpperCAmelCase__ = MusicgenDecoderConfig( hidden_size=lowerCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=lowerCamelCase , num_attention_heads=lowerCamelCase , ) return config @torch.no_grad() def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="cpu" ): UpperCAmelCase__ = MusicGen.get_pretrained(lowerCamelCase , device=lowerCamelCase ) UpperCAmelCase__ = decoder_config_from_checkpoint(lowerCamelCase ) UpperCAmelCase__ = fairseq_model.lm.state_dict() UpperCAmelCase__ , UpperCAmelCase__ = rename_state_dict( lowerCamelCase , hidden_size=decoder_config.hidden_size ) UpperCAmelCase__ = TaEncoderModel.from_pretrained('t5-base' ) UpperCAmelCase__ = EncodecModel.from_pretrained('facebook/encodec_32khz' ) UpperCAmelCase__ = MusicgenForCausalLM(lowerCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase__ , UpperCAmelCase__ = decoder.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(lowerCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model UpperCAmelCase__ = MusicgenForConditionalGeneration(text_encoder=lowerCamelCase , audio_encoder=lowerCamelCase , decoder=lowerCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(lowerCamelCase ) # check we can do a forward pass UpperCAmelCase__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase__ = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits if logits.shape != (8, 1, 2_0_4_8): raise ValueError('Incorrect shape for logits' ) # now construct the processor UpperCAmelCase__ = AutoTokenizer.from_pretrained('t5-base' ) UpperCAmelCase__ = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) UpperCAmelCase__ = MusicgenProcessor(feature_extractor=lowerCamelCase , tokenizer=lowerCamelCase ) # set the appropriate bos/pad token ids UpperCAmelCase__ = 2_0_4_8 UpperCAmelCase__ = 2_0_4_8 # set other default generation config params UpperCAmelCase__ = int(3_0 * audio_encoder.config.frame_rate ) UpperCAmelCase__ = True UpperCAmelCase__ = 3.0 if pytorch_dump_folder is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(lowerCamelCase ) processor.save_pretrained(lowerCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(lowerCamelCase ) processor.push_to_hub(lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) lowerCAmelCase__ : List[str] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
632
0
"""simple docstring""" import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate lowerCAmelCase__ : Dict = trt.Logger(trt.Logger.WARNING) lowerCAmelCase__ : Optional[int] = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) lowerCAmelCase__ : Optional[int] = logging.getLogger(__name__) lowerCAmelCase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--onnx_model_path', default=None, type=str, required=True, help='Path to ONNX model: ', ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='The output directory where the model checkpoints and predictions will be written.', ) # Other parameters parser.add_argument( '--tokenizer_name', default='', type=str, required=True, help='Pretrained tokenizer name or path if not the same as model_name', ) parser.add_argument( '--version_2_with_negative', action='store_true', help='If true, the SQuAD examples contain some that do not have an answer.', ) parser.add_argument( '--null_score_diff_threshold', type=float, default=0.0, help='If null_score - best_non_null is greater than the threshold predict null.', ) parser.add_argument( '--max_seq_length', default=384, type=int, help=( 'The maximum total input sequence length after WordPiece tokenization. Sequences ' 'longer than this will be truncated, and sequences shorter than this will be padded.' ), ) parser.add_argument( '--doc_stride', default=128, type=int, help='When splitting up a long document into chunks, how much stride to take between chunks.', ) parser.add_argument('--per_device_eval_batch_size', default=8, type=int, help='Batch size per GPU/CPU for evaluation.') parser.add_argument( '--n_best_size', default=20, type=int, help='The total number of n-best predictions to generate in the nbest_predictions.json output file.', ) parser.add_argument( '--max_answer_length', default=30, type=int, help=( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ), ) parser.add_argument('--seed', type=int, default=42, help='random seed for initialization') parser.add_argument( '--dataset_name', type=str, default=None, required=True, help='The name of the dataset to use (via the datasets library).', ) parser.add_argument( '--dataset_config_name', type=str, default=None, help='The configuration name of the dataset to use (via the datasets library).', ) parser.add_argument( '--preprocessing_num_workers', type=int, default=4, help='A csv or a json file containing the training data.' ) parser.add_argument('--overwrite_cache', action='store_true', help='Overwrite the cached training and evaluation sets') parser.add_argument( '--fp16', action='store_true', help='Whether to use 16-bit (mixed) precision instead of 32-bit', ) parser.add_argument( '--int8', action='store_true', help='Whether to use INT8', ) lowerCAmelCase__ : Optional[Any] = parser.parse_args() if args.tokenizer_name: lowerCAmelCase__ : str = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported by this script.' 'You can do it from another script, save it, and load it from here, using --tokenizer_name.' ) logger.info('Training/evaluation parameters %s', args) lowerCAmelCase__ : Dict = args.per_device_eval_batch_size lowerCAmelCase__ : Optional[int] = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties lowerCAmelCase__ : int = True lowerCAmelCase__ : Tuple = 'temp_engine/bert-fp32.engine' if args.fpaa: lowerCAmelCase__ : int = 'temp_engine/bert-fp16.engine' if args.inta: lowerCAmelCase__ : Any = 'temp_engine/bert-int8.engine' # import ONNX file if not os.path.exists('temp_engine'): os.makedirs('temp_engine') lowerCAmelCase__ : List[Any] = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, 'rb') as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network lowerCAmelCase__ : Optional[Any] = [network.get_input(i) for i in range(network.num_inputs)] lowerCAmelCase__ : Any = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: lowerCAmelCase__ : Union[str, Any] = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) lowerCAmelCase__ : Tuple = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) lowerCAmelCase__ : Any = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, 'wb') as f: f.write(engine.serialize()) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = np.asarray(inputs['input_ids'] , dtype=np.intaa ) UpperCAmelCase__ = np.asarray(inputs['attention_mask'] , dtype=np.intaa ) UpperCAmelCase__ = np.asarray(inputs['token_type_ids'] , dtype=np.intaa ) # Copy inputs cuda.memcpy_htod_async(d_inputs[0] , input_ids.ravel() , lowerCamelCase ) cuda.memcpy_htod_async(d_inputs[1] , attention_mask.ravel() , lowerCamelCase ) cuda.memcpy_htod_async(d_inputs[2] , token_type_ids.ravel() , lowerCamelCase ) # start time UpperCAmelCase__ = time.time() # Run inference context.execute_async( bindings=[int(lowerCamelCase ) for d_inp in d_inputs] + [int(lowerCamelCase ), int(lowerCamelCase )] , stream_handle=stream.handle ) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(lowerCamelCase , lowerCamelCase , lowerCamelCase ) cuda.memcpy_dtoh_async(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Synchronize the stream and take time stream.synchronize() # end time UpperCAmelCase__ = time.time() UpperCAmelCase__ = end_time - start_time UpperCAmelCase__ = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. lowerCAmelCase__ : List[str] = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. lowerCAmelCase__ : Tuple = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError('Evaluation requires a dataset name') # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. lowerCAmelCase__ : Tuple = raw_datasets['validation'].column_names lowerCAmelCase__ : Optional[Any] = 'question' if 'question' in column_names else column_names[0] lowerCAmelCase__ : Tuple = 'context' if 'context' in column_names else column_names[1] lowerCAmelCase__ : List[str] = 'answers' if 'answers' in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). lowerCAmelCase__ : Any = tokenizer.padding_side == 'right' if args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) lowerCAmelCase__ : Dict = min(args.max_seq_length, tokenizer.model_max_length) def a_ ( lowerCamelCase ): # Some of the questions have lots of whitespace on the left, which is not useful and will make the # truncation of the context fail (the tokenized question will take a lots of space). So we remove that # left whitespace UpperCAmelCase__ = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. UpperCAmelCase__ = tokenizer( examples[question_column_name if pad_on_right else context_column_name] , examples[context_column_name if pad_on_right else question_column_name] , truncation='only_second' if pad_on_right else 'only_first' , max_length=lowerCamelCase , stride=args.doc_stride , return_overflowing_tokens=lowerCamelCase , return_offsets_mapping=lowerCamelCase , padding='max_length' , ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. UpperCAmelCase__ = tokenized_examples.pop('overflow_to_sample_mapping' ) # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. UpperCAmelCase__ = [] for i in range(len(tokenized_examples['input_ids'] ) ): # Grab the sequence corresponding to that example (to know what is the context and what is the question). UpperCAmelCase__ = tokenized_examples.sequence_ids(lowerCamelCase ) UpperCAmelCase__ = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. UpperCAmelCase__ = sample_mapping[i] tokenized_examples["example_id"].append(examples['id'][sample_index] ) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. UpperCAmelCase__ = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples['offset_mapping'][i] ) ] return tokenized_examples lowerCAmelCase__ : Optional[int] = raw_datasets['validation'] # Validation Feature Creation lowerCAmelCase__ : Tuple = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc='Running tokenizer on validation dataset', ) lowerCAmelCase__ : str = default_data_collator lowerCAmelCase__ : Union[str, Any] = eval_dataset.remove_columns(['example_id', 'offset_mapping']) lowerCAmelCase__ : Dict = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase="eval" ): # Post-processing: we match the start logits and end logits to answers in the original context. UpperCAmelCase__ = postprocess_qa_predictions( examples=lowerCamelCase , features=lowerCamelCase , predictions=lowerCamelCase , version_2_with_negative=args.version_2_with_negative , n_best_size=args.n_best_size , max_answer_length=args.max_answer_length , null_score_diff_threshold=args.null_score_diff_threshold , output_dir=args.output_dir , prefix=lowerCamelCase , ) # Format the result to the format the metric expects. if args.version_2_with_negative: UpperCAmelCase__ = [ {'id': k, 'prediction_text': v, 'no_answer_probability': 0.0} for k, v in predictions.items() ] else: UpperCAmelCase__ = [{'id': k, 'prediction_text': v} for k, v in predictions.items()] UpperCAmelCase__ = [{'id': ex['id'], 'answers': ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=lowerCamelCase , label_ids=lowerCamelCase ) lowerCAmelCase__ : str = load_metric('squad_v2' if args.version_2_with_negative else 'squad') # Evaluation! logger.info('Loading ONNX model %s for evaluation', args.onnx_model_path) with open(engine_name, 'rb') as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def a_ ( lowerCamelCase ): return trt.volume(engine.get_binding_shape(lowerCamelCase ) ) * engine.get_binding_dtype(lowerCamelCase ).itemsize # Allocate device memory for inputs and outputs. lowerCAmelCase__ : Tuple = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer lowerCAmelCase__ : Dict = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) lowerCAmelCase__ : Dict = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) lowerCAmelCase__ : Any = cuda.mem_alloc(h_outputa.nbytes) lowerCAmelCase__ : str = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. lowerCAmelCase__ : Any = cuda.Stream() # Evaluation logger.info('***** Running Evaluation *****') logger.info(F""" Num examples = {len(eval_dataset)}""") logger.info(F""" Batch size = {args.per_device_eval_batch_size}""") lowerCAmelCase__ : Any = 0.0 lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : List[Any] = timeit.default_timer() lowerCAmelCase__ : Dict = None for step, batch in enumerate(eval_dataloader): lowerCAmelCase__ : Union[str, Any] = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 lowerCAmelCase__ : List[Any] = outputs lowerCAmelCase__ : Optional[int] = torch.tensor(start_logits) lowerCAmelCase__ : List[str] = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered lowerCAmelCase__ : str = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-100) lowerCAmelCase__ : Any = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-100) lowerCAmelCase__ : Union[str, Any] = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) lowerCAmelCase__ : Union[str, Any] = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-100) if all_preds is not None: lowerCAmelCase__ : Dict = nested_truncate(all_preds, len(eval_dataset)) lowerCAmelCase__ : int = timeit.default_timer() - start_time logger.info(' Evaluation done in total %f secs (%f sec per example)', evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info('Average Inference Time = {:.3f} ms'.format(total_time * 1_000 / niter)) logger.info('Total Inference Time = {:.3f} ms'.format(total_time * 1_000)) logger.info('Total Number of Inference = %d', niter) lowerCAmelCase__ : Tuple = post_processing_function(eval_examples, eval_dataset, all_preds) lowerCAmelCase__ : Dict = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(F"""Evaluation metrics: {eval_metric}""")
710
"""simple docstring""" lowerCAmelCase__ : Tuple = range(2, 20 + 1) lowerCAmelCase__ : Optional[Any] = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase__ : dict[int, dict[int, list[list[int]]]] = {} def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(a_i[j] for j in range(lowerCamelCase , len(lowerCamelCase ) ) ) UpperCAmelCase__ = sum(a_i[j] * base[j] for j in range(min(len(lowerCamelCase ) , lowerCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ = 0, 0 UpperCAmelCase__ = n - i UpperCAmelCase__ = memo.get(lowerCamelCase ) if sub_memo is not None: UpperCAmelCase__ = sub_memo.get(lowerCamelCase ) if jumps is not None and len(lowerCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ = -1 for _k in range(len(lowerCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ = diff + c for j in range(min(lowerCamelCase , len(lowerCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) if new_c > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: UpperCAmelCase__ = [] else: UpperCAmelCase__ = {c: []} UpperCAmelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , k - 1 , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ = compute(lowerCamelCase , lowerCamelCase , i + dn , lowerCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ = 0 while j < len(lowerCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowerCamelCase , (diff, dn, k) ) return (diff, dn) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if i >= n: return 0, i if k > len(lowerCamelCase ): a_i.extend([0 for _ in range(k - len(lowerCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 0, 0, 0 for j in range(len(lowerCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ = ds_c + ds_b diff += addend UpperCAmelCase__ = 0 for j in range(lowerCamelCase ): UpperCAmelCase__ = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return diff, i - start_i def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): for j in range(lowerCamelCase , len(lowerCamelCase ) ): UpperCAmelCase__ = digits[j] + addend if s >= 1_0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) UpperCAmelCase__ = addend // 1_0 + quotient else: UpperCAmelCase__ = s UpperCAmelCase__ = addend // 1_0 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ = divmod(lowerCamelCase , 1_0 ) digits.append(lowerCamelCase ) def a_ ( lowerCamelCase = 1_0**1_5 ): UpperCAmelCase__ = [1] UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ = next_term(lowerCamelCase , 2_0 , i + dn , lowerCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ = 0 for j in range(len(lowerCamelCase ) ): a_n += digits[j] * 1_0**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
632
0
"""simple docstring""" from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES lowerCAmelCase__ : List[str] = logging.get_logger(__name__) lowerCAmelCase__ : List[str] = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) lowerCAmelCase__ : Any = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) lowerCAmelCase__ : Dict = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) lowerCAmelCase__ : int = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) lowerCAmelCase__ : Optional[Any] = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) lowerCAmelCase__ : str = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) lowerCAmelCase__ : Any = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) lowerCAmelCase__ : Any = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) lowerCAmelCase__ : int = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) lowerCAmelCase__ : Tuple = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) lowerCAmelCase__ : Tuple = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) lowerCAmelCase__ : Tuple = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) lowerCAmelCase__ : int = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) lowerCAmelCase__ : Any = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) lowerCAmelCase__ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) lowerCAmelCase__ : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) lowerCAmelCase__ : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) lowerCAmelCase__ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) lowerCAmelCase__ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) lowerCAmelCase__ : Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) lowerCAmelCase__ : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) lowerCAmelCase__ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) lowerCAmelCase__ : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) lowerCAmelCase__ : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) lowerCAmelCase__ : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) lowerCAmelCase__ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) lowerCAmelCase__ : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) lowerCAmelCase__ : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_MAPPING lowerCAmelCase__ : Optional[int] = auto_class_update(FlaxAutoModel) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_PRETRAINING_MAPPING lowerCAmelCase__ : Optional[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING lowerCAmelCase__ : Optional[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_MASKED_LM_MAPPING lowerCAmelCase__ : List[str] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCAmelCase__ : Dict = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowerCAmelCase__ : int = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING lowerCAmelCase__ : Optional[int] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCAmelCase__ : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING lowerCAmelCase__ : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING lowerCAmelCase__ : Optional[int] = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCAmelCase__ : List[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING lowerCAmelCase__ : Union[str, Any] = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class snake_case ( _BaseAutoModelClass ): """simple docstring""" snake_case__ = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING lowerCAmelCase__ : Optional[int] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
711
"""simple docstring""" import random class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : str ): UpperCAmelCase__ = [ord(lowerCamelCase__ ) for i in text] UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i in plain: UpperCAmelCase__ = random.randint(1 ,300 ) UpperCAmelCase__ = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def __lowerCAmelCase ( lowerCamelCase__ : list[int] ,lowerCamelCase__ : list[int] ): UpperCAmelCase__ = [] for i in range(len(lowerCamelCase__ ) ): UpperCAmelCase__ = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ : Dict = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
632
0
"""simple docstring""" from __future__ import annotations def a_ ( lowerCamelCase ): UpperCAmelCase__ = str(lowerCamelCase ) return len(lowerCamelCase ) == 9 and set(lowerCamelCase ) == set('123456789' ) def a_ ( ): for base_num in range(9_9_9_9 , 4_9_9_9 , -1 ): UpperCAmelCase__ = 1_0_0_0_0_2 * base_num if is_9_pandigital(lowerCamelCase ): return candidate for base_num in range(3_3_3 , 9_9 , -1 ): UpperCAmelCase__ = 1_0_0_2_0_0_3 * base_num if is_9_pandigital(lowerCamelCase ): return candidate return None if __name__ == "__main__": print(F"""{solution() = }""")
712
"""simple docstring""" import re def a_ ( lowerCamelCase ): return [char.split() for char in re.split(r'[^ a-z A-Z 0-9 \s]' , str_ )] def a_ ( lowerCamelCase ): UpperCAmelCase__ = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): try: UpperCAmelCase__ = split_input(lowerCamelCase ) if upper: UpperCAmelCase__ = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: UpperCAmelCase__ = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCamelCase ): return to_simple_case(lowerCamelCase ) def a_ ( lowerCamelCase ): try: UpperCAmelCase__ = to_simple_case(lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '_' ) def a_ ( lowerCamelCase , lowerCamelCase ): return to_complex_case(lowerCamelCase , lowerCamelCase , '-' ) if __name__ == "__main__": __import__('doctest').testmod()
632
0
"""simple docstring""" def a_ ( lowerCamelCase ): UpperCAmelCase__ = 1 for i in range(1 , num + 1 ): fact *= i return fact def a_ ( lowerCamelCase ): UpperCAmelCase__ = 0 while number > 0: UpperCAmelCase__ = number % 1_0 sum_of_digits += last_digit UpperCAmelCase__ = number // 1_0 # Removing the last_digit from the given number return sum_of_digits def a_ ( lowerCamelCase = 1_0_0 ): UpperCAmelCase__ = factorial(lowerCamelCase ) UpperCAmelCase__ = split_and_add(lowerCamelCase ) return result if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
713
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Dict ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_torch def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,torch.tensor(lowerCamelCase__ ) ,torch.tensor(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ) @require_vision @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : str ,**lowerCamelCase__ : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = self.get_image_processor(do_normalize=lowerCamelCase__ ,padding_value=1.0 ) UpperCAmelCase__ = SamProcessor.from_pretrained(self.tmpdirname ,do_normalize=lowerCamelCase__ ,padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='np' ) UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='np' ) input_feat_extract.pop('original_sizes' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) @require_tf def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,tf.convert_to_tensor(lowerCamelCase__ ) ,tf.convert_to_tensor(lowerCamelCase__ ) ,return_tensors='tf' ,) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) # should also work with np UpperCAmelCase__ = [np.ones((1, 3, 5, 5) )] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) self.assertEqual(masks[0].shape ,(1, 3, 1_764, 2_646) ) UpperCAmelCase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,np.array(lowerCamelCase__ ) ,np.array(lowerCamelCase__ ) ,return_tensors='tf' ) @require_vision @require_torchvision class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = tempfile.mkdtemp() UpperCAmelCase__ = SamImageProcessor() UpperCAmelCase__ = SamProcessor(lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowerCamelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ).image_processor def __lowerCAmelCase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase__ = [Image.fromarray(np.moveaxis(lowerCamelCase__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = np.random.randint(0 ,2 ,size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase__ = [tf.convert_to_tensor(lowerCamelCase__ )] UpperCAmelCase__ = [torch.tensor(lowerCamelCase__ )] UpperCAmelCase__ = [[1_764, 2_646]] UpperCAmelCase__ = [[683, 1_024]] UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='tf' ) UpperCAmelCase__ = processor.post_process_masks( lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,return_tensors='pt' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = SamProcessor(image_processor=lowerCamelCase__ ) UpperCAmelCase__ = self.prepare_image_inputs() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='pt' )['pixel_values'].numpy() UpperCAmelCase__ = image_processor(lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() UpperCAmelCase__ = processor(images=lowerCamelCase__ ,return_tensors='tf' )['pixel_values'].numpy() self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ) )
632
0