code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import math def lowerCAmelCase_ ( _lowercase : Union[str, Any]) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_lowercase) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCAmelCase_ ( _lowercase : str = 0.1) -> int: """simple docstring""" a__ : Tuple = 3 a__ : List[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1): primes += is_prime(_lowercase) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
136
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): """simple docstring""" lowercase : int = LongformerTokenizer lowercase : List[str] = True lowercase : int = LongformerTokenizerFast lowercase : Optional[Any] = True def __lowerCamelCase ( self ) -> List[Any]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase : Optional[int] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase : List[str] = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __UpperCamelCase : Union[str, Any] = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase : Optional[int] = {'unk_token': '<unk>'} __UpperCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __UpperCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_lowerCAmelCase ) ) def __lowerCamelCase ( self , **__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __lowerCamelCase ( self , **__UpperCamelCase ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __lowerCamelCase ( self , __UpperCamelCase ) -> List[str]: '''simple docstring''' __UpperCamelCase : Dict = 'lower newer' __UpperCamelCase : Union[str, Any] = 'lower newer' return input_text, output_text def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase : List[Any] = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __UpperCamelCase : str = 'lower newer' __UpperCamelCase : List[Any] = ['l', 'o', 'w', 'er', '\u0120', 'n', 'e', 'w', 'er'] __UpperCamelCase : Optional[int] = tokenizer.tokenize(_lowerCAmelCase ) # , add_prefix_space=True) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : Dict = tokens + [tokenizer.unk_token] __UpperCamelCase : Optional[Any] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , _lowerCAmelCase ) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' __UpperCamelCase : List[Any] = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" , add_special_tokens=_lowerCAmelCase ) , [0, 3_14_14, 2_32, 3_28, 2] ) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" , add_special_tokens=_lowerCAmelCase ) , [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2] , ) @slow def __lowerCamelCase ( self ) -> Tuple: '''simple docstring''' __UpperCamelCase : Any = self.tokenizer_class.from_pretrained("allenai/longformer-base-4096" ) __UpperCamelCase : Optional[Any] = tokenizer.encode("sequence builders" , add_special_tokens=_lowerCAmelCase ) __UpperCamelCase : str = tokenizer.encode("multi-sequence build" , add_special_tokens=_lowerCAmelCase ) __UpperCamelCase : Any = tokenizer.encode( "sequence builders" , add_special_tokens=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) __UpperCamelCase : List[str] = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) __UpperCamelCase : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) __UpperCamelCase : Dict = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : List[str] = self.get_tokenizer() __UpperCamelCase : Tuple = 'Encode this sequence.' __UpperCamelCase : List[Any] = tokenizer.byte_encoder[' '.encode("utf-8" )[0]] # Testing encoder arguments __UpperCamelCase : int = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) __UpperCamelCase : Any = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(_lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : Dict = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) __UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) tokenizer.add_special_tokens({"bos_token": "<s>"} ) __UpperCamelCase : Optional[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) __UpperCamelCase : Dict = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(_lowerCAmelCase , _lowerCAmelCase ) # Testing spaces after special tokens __UpperCamelCase : int = '<mask>' tokenizer.add_special_tokens( {"mask_token": AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase )} ) # mask token has a left space __UpperCamelCase : int = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) __UpperCamelCase : List[Any] = 'Encode <mask> sequence' __UpperCamelCase : List[Any] = 'Encode <mask>sequence' __UpperCamelCase : Union[str, Any] = tokenizer.encode(_lowerCAmelCase ) __UpperCamelCase : int = encoded.index(_lowerCAmelCase ) __UpperCamelCase : Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : Tuple = tokenizer.encode(_lowerCAmelCase ) __UpperCamelCase : Optional[Any] = encoded.index(_lowerCAmelCase ) __UpperCamelCase : int = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(_lowerCAmelCase , _lowerCAmelCase ) def __lowerCamelCase ( self ) -> Optional[Any]: '''simple docstring''' pass def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCamelCase : Dict = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __UpperCamelCase : Dict = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __UpperCamelCase : List[str] = 'A, <mask> AllenNLP sentence.' __UpperCamelCase : Any = tokenizer_r.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase ) __UpperCamelCase : Any = tokenizer_p.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) __UpperCamelCase : str = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) __UpperCamelCase : Any = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( _lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( _lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __UpperCamelCase : Any = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : str = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __UpperCamelCase : Optional[Any] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["add_prefix_space"] , _lowerCAmelCase ) self.assertEqual(post_processor_state["add_prefix_space"] , _lowerCAmelCase ) self.assertEqual(post_processor_state["trim_offsets"] , _lowerCAmelCase ) def __lowerCamelCase ( self ) -> List[str]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCamelCase : int = 'hello' # `hello` is a token in the vocabulary of `pretrained_name` __UpperCamelCase : str = f'''{text_of_1_token} {text_of_1_token}''' __UpperCamelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : Optional[int] = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCAmelCase ) + 1, len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : str = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : Union[str, Any] = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCAmelCase ) + 1, len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : Optional[Any] = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : Dict = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCAmelCase ), len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : int = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : int = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_lowerCAmelCase ), len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : Optional[int] = f''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __UpperCamelCase : Optional[Any] = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : Optional[Any] = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCAmelCase ) + 1, 1 + len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : List[Any] = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : List[Any] = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCAmelCase ), 1 + len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , ) __UpperCamelCase : Tuple = self.rust_tokenizer_class.from_pretrained( _lowerCAmelCase , use_fast=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase ) __UpperCamelCase : Union[str, Any] = tokenizer_r(_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_lowerCAmelCase ), 1 + len(_lowerCAmelCase ) + 1 + len(_lowerCAmelCase )) , )
327
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
0
'''simple docstring''' from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class _lowerCAmelCase : def __init__(self , lowercase , ): A_ : List[Any] = parent A_ : List[str] = 13 A_ : str = 7 A_ : int = 30 A_ : Optional[Any] = self.seq_length + self.mem_len A_ : List[str] = 15 A_ : Tuple = True A_ : Optional[Any] = True A_ : List[Any] = 99 A_ : int = [10, 50, 80] A_ : Tuple = 32 A_ : Any = 32 A_ : int = 4 A_ : List[str] = 8 A_ : Tuple = 128 A_ : int = 2 A_ : Optional[Any] = 2 A_ : str = None A_ : int = 1 A_ : Tuple = 0 A_ : Dict = 3 A_ : Union[str, Any] = self.vocab_size - 1 A_ : Optional[Any] = 0.01 def _a (self ): A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : List[Any] = None if self.use_labels: A_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Any = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def _a (self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _a (self , lowercase , lowercase , lowercase , lowercase ): A_ : Union[str, Any] = TFTransfoXLModel(_lowerCAmelCase ) A_ : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() A_ : Any = {'input_ids': input_ids_a, 'mems': mems_a} A_ : Any = model(_lowerCAmelCase ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _a (self , lowercase , lowercase , lowercase , lowercase ): A_ : Dict = TFTransfoXLLMHeadModel(_lowerCAmelCase ) A_ : List[Any] = model(_lowerCAmelCase ).to_tuple() A_ : Optional[int] = {'input_ids': input_ids_a, 'labels': lm_labels} A_ : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() A_ : Dict = model([input_ids_a, mems_a] ).to_tuple() A_ : Any = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} A_ : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _a (self , lowercase , lowercase , lowercase , lowercase ): A_ : str = TFTransfoXLForSequenceClassification(_lowerCAmelCase ) A_ : Dict = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a (self ): A_ : Any = self.prepare_config_and_inputs() (A_) : str = config_and_inputs A_ : List[Any] = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): __SCREAMING_SNAKE_CASE : List[str] = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) __SCREAMING_SNAKE_CASE : Optional[Any] = () if is_tf_available() else () __SCREAMING_SNAKE_CASE : Any = ( { "feature-extraction": TFTransfoXLModel, "text-classification": TFTransfoXLForSequenceClassification, "text-generation": TFTransfoXLLMHeadModel, "zero-shot": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented __SCREAMING_SNAKE_CASE : int = False __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : str = False __SCREAMING_SNAKE_CASE : Optional[Any] = False def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _a (self ): A_ : Optional[Any] = TFTransfoXLModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=_lowerCAmelCase , d_embed=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): self.model_tester.set_seed() A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*_lowerCAmelCase ) def _a (self ): self.model_tester.set_seed() A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*_lowerCAmelCase ) def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*_lowerCAmelCase ) def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: A_ : int = model_class(_lowerCAmelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: A_ : Dict = model.get_output_embeddings() assert isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) A_ : List[Any] = model.get_bias() assert name is None else: A_ : List[Any] = model.get_output_embeddings() assert x is None A_ : int = model.get_bias() assert name is None def _a (self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _a (self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Union[str, Any] = TFTransfoXLModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip(reason="""This model doesn\'t play well with fit() due to not returning a single loss.""" ) def _a (self ): pass @require_tf class _lowerCAmelCase ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _a (self ): A_ : str = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off A_ : Union[str, Any] = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off A_ : Optional[int] = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> A_ : Union[str, Any] = model.generate(_lowerCAmelCase , max_length=200 , do_sample=_lowerCAmelCase ) self.assertListEqual(output_ids[0].numpy().tolist() , _lowerCAmelCase )
667
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
0
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): @slow def lowercase__ ( self : Any ): lowerCAmelCase : str = TFXLMRobertaModel.from_pretrained('jplu/tf-xlm-roberta-base' ) lowerCAmelCase : Tuple = { 'input_ids': tf.convert_to_tensor([[0, 2646, 10269, 83, 99942, 2]] , dtype=tf.intaa ), # "My dog is cute" 'attention_mask': tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } lowerCAmelCase : Optional[int] = model(_lowerCAmelCase )['last_hidden_state'] lowerCAmelCase : Union[str, Any] = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) # compare the actual values for a slice. lowerCAmelCase : List[Any] = tf.convert_to_tensor( [ [ [0.0_68_17_62, 0.10_89_44_51, 0.06_77_25_04], [-0.06_42_36_68, 0.02_36_66_15, 0.04_32_93_44], [-0.06_05_72_95, 0.09_97_41_35, -0.00_07_05_84], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
343
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
0
import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset __lowerCAmelCase = 'bert-base-cased' __lowerCAmelCase = 'google/pegasus-xsum' __lowerCAmelCase = [' Sam ate lunch today.', 'Sams lunch ingredients.'] __lowerCAmelCase = ['A very interesting story about what I ate for lunch.', 'Avocado, celery, turkey, coffee'] __lowerCAmelCase = 'patrickvonplaten/t5-tiny-random' __lowerCAmelCase = 'sshleifer/bart-tiny-random' __lowerCAmelCase = 'sshleifer/tiny-mbart' __lowerCAmelCase = 'sshleifer/tiny-marian-en-de' def a ( a , a ) ->int: '''simple docstring''' SCREAMING_SNAKE_CASE = '\n'.join(a ) Path(a ).open('''w''' ).writelines(a ) def a ( a ) ->Optional[int]: '''simple docstring''' for split in ["train", "val", "test"]: _dump_articles(os.path.join(a , F"""{split}.source""" ) , a ) _dump_articles(os.path.join(a , F"""{split}.target""" ) , a ) return tmp_dir class lowerCamelCase ( __snake_case ): @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def snake_case__ ( self :Optional[Any] , lowercase :Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) SCREAMING_SNAKE_CASE = max(len(tokenizer.encode(_lowerCAmelCase ) ) for a in ARTICLES ) SCREAMING_SNAKE_CASE = max(len(tokenizer.encode(_lowerCAmelCase ) ) for a in SUMMARIES ) SCREAMING_SNAKE_CASE = 4 SCREAMING_SNAKE_CASE = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated SCREAMING_SNAKE_CASE = 'ro_RO', 'de_DE' # ignored for all but mbart, but never causes error. SCREAMING_SNAKE_CASE = SeqaSeqDataset( _lowerCAmelCase , data_dir=_lowerCAmelCase , type_path='''train''' , max_source_length=_lowerCAmelCase , max_target_length=_lowerCAmelCase , src_lang=_lowerCAmelCase , tgt_lang=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE = DataLoader(_lowerCAmelCase , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place SCREAMING_SNAKE_CASE = shift_tokens_right(batch['''labels'''] , tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def snake_case__ ( self :List[Any] , lowercase :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) SCREAMING_SNAKE_CASE = max(len(tokenizer.encode(_lowerCAmelCase ) ) for a in ARTICLES ) SCREAMING_SNAKE_CASE = max(len(tokenizer.encode(_lowerCAmelCase ) ) for a in SUMMARIES ) SCREAMING_SNAKE_CASE = 4 SCREAMING_SNAKE_CASE = LegacySeqaSeqDataset( _lowerCAmelCase , data_dir=_lowerCAmelCase , type_path='''train''' , max_source_length=2_0 , max_target_length=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE = DataLoader(_lowerCAmelCase , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 2_0 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def snake_case__ ( self :List[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained('''facebook/mbart-large-cc25''' ) SCREAMING_SNAKE_CASE = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) SCREAMING_SNAKE_CASE = tmp_dir.joinpath('''train.source''' ).open().readlines() SCREAMING_SNAKE_CASE = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(_lowerCAmelCase , _lowerCAmelCase , 1_2_8 , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = {x.name for x in tmp_dir.iterdir()} SCREAMING_SNAKE_CASE = {x.name for x in save_dir.iterdir()} SCREAMING_SNAKE_CASE = save_dir.joinpath('''train.source''' ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(_lowerCAmelCase ) < len(_lowerCAmelCase ) assert len(_lowerCAmelCase ) == 1 assert len(packed_examples[0] ) == sum(len(_lowerCAmelCase ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason='''This test requires fairseq''' ) def snake_case__ ( self :int ) -> Optional[Any]: """simple docstring""" if not FAIRSEQ_AVAILABLE: return SCREAMING_SNAKE_CASE = self._get_dataset(max_len=6_4 ) SCREAMING_SNAKE_CASE = 6_4 SCREAMING_SNAKE_CASE = ds.make_dynamic_sampler(_lowerCAmelCase , required_batch_size_multiple=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = [len(_lowerCAmelCase ) for x in batch_sampler] assert len(set(_lowerCAmelCase ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(_lowerCAmelCase ) == len(_lowerCAmelCase ) # no dropped or added examples SCREAMING_SNAKE_CASE = DataLoader(_lowerCAmelCase , batch_sampler=_lowerCAmelCase , collate_fn=ds.collate_fn , num_workers=2 ) SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for batch in data_loader: SCREAMING_SNAKE_CASE = batch['input_ids'].shape SCREAMING_SNAKE_CASE = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple SCREAMING_SNAKE_CASE = np.product(batch['''input_ids'''].shape ) num_src_per_batch.append(_lowerCAmelCase ) if num_src_tokens > (max_tokens * 1.1): failures.append(_lowerCAmelCase ) assert num_src_per_batch[0] == max(_lowerCAmelCase ) if failures: raise AssertionError(f"""too many tokens in {len(_lowerCAmelCase )} batches""" ) def snake_case__ ( self :Optional[int] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE = self._get_dataset(max_len=5_1_2 ) SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = ds.make_sortish_sampler(_lowerCAmelCase , shuffle=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = DataLoader(_lowerCAmelCase , batch_size=_lowerCAmelCase , collate_fn=ds.collate_fn , num_workers=2 ) SCREAMING_SNAKE_CASE = DataLoader(_lowerCAmelCase , batch_size=_lowerCAmelCase , collate_fn=ds.collate_fn , num_workers=2 , sampler=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = tokenizer.pad_token_id def count_pad_tokens(lowercase :Any , lowercase :Any="input_ids" ): return [batch[k].eq(_lowerCAmelCase ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(_lowerCAmelCase , k='''labels''' ) ) < sum(count_pad_tokens(_lowerCAmelCase , k='''labels''' ) ) assert sum(count_pad_tokens(_lowerCAmelCase ) ) < sum(count_pad_tokens(_lowerCAmelCase ) ) assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) def snake_case__ ( self :Optional[Any] , lowercase :List[Any]=1_0_0_0 , lowercase :Dict=1_2_8 ) -> Tuple: """simple docstring""" if os.getenv('''USE_REAL_DATA''' , _lowerCAmelCase ): SCREAMING_SNAKE_CASE = 'examples/seq2seq/wmt_en_ro' SCREAMING_SNAKE_CASE = max_len * 2 * 6_4 if not Path(_lowerCAmelCase ).joinpath('''train.len''' ).exists(): save_len_file(_lowerCAmelCase , _lowerCAmelCase ) else: SCREAMING_SNAKE_CASE = 'examples/seq2seq/test_data/wmt_en_ro' SCREAMING_SNAKE_CASE = max_len * 4 save_len_file(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = SeqaSeqDataset( _lowerCAmelCase , data_dir=_lowerCAmelCase , type_path='''train''' , max_source_length=_lowerCAmelCase , max_target_length=_lowerCAmelCase , n_obs=_lowerCAmelCase , ) return ds, max_tokens, tokenizer def snake_case__ ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE = self._get_dataset() SCREAMING_SNAKE_CASE = set(DistributedSortishSampler(_lowerCAmelCase , 2_5_6 , num_replicas=2 , rank=0 , add_extra_examples=_lowerCAmelCase ) ) SCREAMING_SNAKE_CASE = set(DistributedSortishSampler(_lowerCAmelCase , 2_5_6 , num_replicas=2 , rank=1 , add_extra_examples=_lowerCAmelCase ) ) assert idsa.intersection(_lowerCAmelCase ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def snake_case__ ( self :Any , lowercase :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_lowerCAmelCase , use_fast=_lowerCAmelCase ) if tok_name == MBART_TINY: SCREAMING_SNAKE_CASE = SeqaSeqDataset( _lowerCAmelCase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , src_lang='''EN''' , tgt_lang='''FR''' , ) SCREAMING_SNAKE_CASE = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: SCREAMING_SNAKE_CASE = SeqaSeqDataset( _lowerCAmelCase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , ) SCREAMING_SNAKE_CASE = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(_lowerCAmelCase ) == 1 if tok_name == BART_TINY else len(_lowerCAmelCase ) == 0
201
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE :List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE :Any = { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/config.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/config.json''' # See all FNet models at https://huggingface.co/models?filter=fnet } class A_ ( __snake_case ): _lowerCamelCase : Optional[int] = "fnet" def __init__( self : str , snake_case_ : Optional[int]=3_2_0_0_0 , snake_case_ : Optional[int]=7_6_8 , snake_case_ : List[Any]=1_2 , snake_case_ : int=3_0_7_2 , snake_case_ : int="gelu_new" , snake_case_ : Dict=0.1 , snake_case_ : str=5_1_2 , snake_case_ : int=4 , snake_case_ : int=0.0_2 , snake_case_ : List[Any]=1e-12 , snake_case_ : Any=False , snake_case_ : List[Any]=5_1_2 , snake_case_ : Union[str, Any]=3 , snake_case_ : Dict=1 , snake_case_ : Optional[Any]=2 , **snake_case_ : List[str] , ): super().__init__(pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , **_lowerCAmelCase ) _UpperCAmelCase = vocab_size _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = initializer_range _UpperCAmelCase = type_vocab_size _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = use_tpu_fourier_optimizations _UpperCAmelCase = tpu_short_seq_length
236
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
0
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging __A =logging.get_logger(__name__) def lowerCamelCase_ ( lowerCamelCase__=None , lowerCamelCase__=None ): return field(default_factory=lambda: default , metadata=lowerCamelCase__ ) @dataclass class _SCREAMING_SNAKE_CASE : lowerCAmelCase__ = list_field( default=[] , metadata={ 'help': ( 'Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version' ' of all available models' ) } , ) lowerCAmelCase__ = list_field( default=[8] , metadata={'help': 'List of batch sizes for which memory and time performance will be evaluated'} ) lowerCAmelCase__ = list_field( default=[8, 32, 1_28, 5_12] , metadata={'help': 'List of sequence lengths for which memory and time performance will be evaluated'} , ) lowerCAmelCase__ = field( default=__snake_case , metadata={'help': 'Whether to benchmark inference of model. Inference can be disabled via --no-inference.'} , ) lowerCAmelCase__ = field( default=__snake_case , metadata={'help': 'Whether to run on available cuda devices. Cuda can be disabled via --no-cuda.'} , ) lowerCAmelCase__ = field( default=__snake_case , metadata={'help': 'Whether to run on available tpu devices. TPU can be disabled via --no-tpu.'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Use FP16 to accelerate inference.'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Benchmark training of model'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Verbose memory tracing'} ) lowerCAmelCase__ = field( default=__snake_case , metadata={'help': 'Whether to perform speed measurements. Speed measurements can be disabled via --no-speed.'} , ) lowerCAmelCase__ = field( default=__snake_case , metadata={ 'help': 'Whether to perform memory measurements. Memory measurements can be disabled via --no-memory' } , ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Trace memory line by line'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Save result to a CSV file'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Save all print statements in a log file'} ) lowerCAmelCase__ = field(default=__snake_case , metadata={'help': 'Whether to print environment information'} ) lowerCAmelCase__ = field( default=__snake_case , metadata={ 'help': ( 'Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use' ' multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled' ' for debugging / testing and on TPU.' ) } , ) lowerCAmelCase__ = field( default=F"""inference_time_{round(time() )}.csv""" , metadata={'help': 'CSV filename used if saving time results to csv.'} , ) lowerCAmelCase__ = field( default=F"""inference_memory_{round(time() )}.csv""" , metadata={'help': 'CSV filename used if saving memory results to csv.'} , ) lowerCAmelCase__ = field( default=F"""train_time_{round(time() )}.csv""" , metadata={'help': 'CSV filename used if saving time results to csv for training.'} , ) lowerCAmelCase__ = field( default=F"""train_memory_{round(time() )}.csv""" , metadata={'help': 'CSV filename used if saving memory results to csv for training.'} , ) lowerCAmelCase__ = field( default=F"""env_info_{round(time() )}.csv""" , metadata={'help': 'CSV filename used if saving environment information.'} , ) lowerCAmelCase__ = field( default=F"""log_{round(time() )}.csv""" , metadata={'help': 'Log filename used if print statements are saved in log.'} , ) lowerCAmelCase__ = field(default=3 , metadata={'help': 'Times an experiment will be run.'} ) lowerCAmelCase__ = field( default=__snake_case , metadata={ 'help': ( 'Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain' ' model weights.' ) } , ) def SCREAMING_SNAKE_CASE_( self ) -> Any: warnings.warn( f'The class {self.__class__} is deprecated. Hugging Face Benchmarking utils' " are deprecated in general and it is advised to use external Benchmarking libraries " " to benchmark Transformer models." , _lowerCAmelCase , ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: if len(self.models ) <= 0: raise ValueError( "Please make sure you provide at least one model name / model identifier, *e.g.* `--models" " bert-base-cased` or `args.models = [\'bert-base-cased\']." ) return self.models @property def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: if not self.multi_process: return False elif self.is_tpu: logger.info("Multiprocessing is currently not possible on TPU." ) return False else: return True
463
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
0
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input __A : str = '''Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine''' def A_ ( ): '''simple docstring''' UpperCamelCase : Dict = _ask_options( """In which compute environment are you running?""" ,["""This machine""", """AWS (Amazon SageMaker)"""] ,_convert_compute_environment ,) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCamelCase : Union[str, Any] = get_sagemaker_input() else: UpperCamelCase : str = get_cluster_input() return config def A_ ( snake_case_ : Any=None ): '''simple docstring''' if subparsers is not None: UpperCamelCase : Union[str, Any] = subparsers.add_parser("""config""" ,description=snake_case_ ) else: UpperCamelCase : List[str] = argparse.ArgumentParser("""Accelerate config command""" ,description=snake_case_ ) parser.add_argument( """--config_file""" ,default=snake_case_ ,help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have """ """such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed """ """with \'huggingface\'.""" ) ,) if subparsers is not None: parser.set_defaults(func=snake_case_ ) return parser def A_ ( snake_case_ : Optional[Any] ): '''simple docstring''' UpperCamelCase : List[str] = get_user_input() if args.config_file is not None: UpperCamelCase : Optional[Any] = args.config_file else: if not os.path.isdir(snake_case_ ): os.makedirs(snake_case_ ) UpperCamelCase : List[str] = default_yaml_config_file if config_file.endswith(""".json""" ): config.to_json_file(snake_case_ ) else: config.to_yaml_file(snake_case_ ) print(f'accelerate configuration saved at {config_file}' ) def A_ ( ): '''simple docstring''' UpperCamelCase : Union[str, Any] = config_command_parser() UpperCamelCase : Any = parser.parse_args() config_command(snake_case_ ) if __name__ == "__main__": main()
499
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
0
'''simple docstring''' from __future__ import annotations import numpy as np def __A ( _SCREAMING_SNAKE_CASE : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = np.shape(_SCREAMING_SNAKE_CASE ) if rows != columns: __SCREAMING_SNAKE_CASE : int = ( '\'table\' has to be of square shaped array but got a ' f'{rows}x{columns} array:\n{table}' ) raise ValueError(_SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Optional[Any] = np.zeros((rows, columns) ) __SCREAMING_SNAKE_CASE : Optional[Any] = np.zeros((rows, columns) ) for i in range(_SCREAMING_SNAKE_CASE ): for j in range(_SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Optional[Any] = sum(lower[i][k] * upper[k][j] for k in range(_SCREAMING_SNAKE_CASE ) ) if upper[j][j] == 0: raise ArithmeticError("No LU decomposition exists" ) __SCREAMING_SNAKE_CASE : Dict = (table[i][j] - total) / upper[j][j] __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 for j in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Any = sum(lower[i][k] * upper[k][j] for k in range(_SCREAMING_SNAKE_CASE ) ) __SCREAMING_SNAKE_CASE : Tuple = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
211
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
0
'''simple docstring''' from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class __magic_name__ ( __snake_case ): UpperCamelCase__ = "" UpperCamelCase__ = "hf-legacy" # "hf://"" is reserved for hffs def __init__( self , snake_case_ = None , snake_case_ = None , **snake_case_ , ): super().__init__(self , **_lowerCAmelCase ) lowercase =repo_info lowercase =token lowercase =None def _A( self ): if self.dir_cache is None: lowercase ={} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes lowercase ={ 'name': hf_file.rfilename, 'size': None, 'type': 'file', } self.dir_cache.update( { str(_lowerCAmelCase ): {'''name''': str(_lowerCAmelCase ), '''size''': None, '''type''': '''directory'''} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def _A( self , snake_case_ , snake_case_ = "rb" , **snake_case_ , ): if not isinstance(self.repo_info , _lowerCAmelCase ): raise NotImplementedError(f'Open is only implemented for dataset repositories, but got {self.repo_info}' ) lowercase =hf_hub_url(self.repo_info.id , _lowerCAmelCase , revision=self.repo_info.sha ) return fsspec.open( _lowerCAmelCase , mode=_lowerCAmelCase , headers=get_authentication_headers_for_url(_lowerCAmelCase , use_auth_token=self.token ) , client_kwargs={'''trust_env''': True} , ).open() def _A( self , snake_case_ , **snake_case_ ): self._get_dirs() lowercase =self._strip_protocol(_lowerCAmelCase ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_=False , **snake_case_ ): self._get_dirs() lowercase =PurePosixPath(path.strip('''/''' ) ) lowercase ={} for p, f in self.dir_cache.items(): lowercase =PurePosixPath(p.strip('''/''' ) ) lowercase =p.parent if root == path: lowercase =f lowercase =list(paths.values() ) if detail: return out else: return sorted(f['''name'''] for f in out )
72
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
0
from __future__ import annotations from typing import Any class A ( __snake_case ): pass class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : List[Any] ) -> int: """simple docstring""" UpperCAmelCase__ = data UpperCAmelCase__ = None def __iter__(self : int ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self UpperCAmelCase__ = [] while node: if node in visited: raise ContainsLoopError visited.append(_lowerCAmelCase ) yield node.data UpperCAmelCase__ = node.next_node @property def lowercase_ (self : List[str] ) -> List[Any]: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": UpperCamelCase__ = Node(1) UpperCamelCase__ = Node(2) UpperCamelCase__ = Node(3) UpperCamelCase__ = Node(4) print(root_node.has_loop) # False UpperCamelCase__ = root_node.next_node print(root_node.has_loop) # True UpperCamelCase__ = Node(5) UpperCamelCase__ = Node(6) UpperCamelCase__ = Node(5) UpperCamelCase__ = Node(6) print(root_node.has_loop) # False UpperCamelCase__ = Node(1) print(root_node.has_loop) # False
486
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
0
import sys _lowercase : Optional[int] =( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def lowerCAmelCase_ ( _lowercase : Optional[Any]) -> int: """simple docstring""" a__ : List[Any] = 1 for digit in s: product *= int(_lowercase) return product def lowerCAmelCase_ ( _lowercase : Optional[int] = N) -> int: """simple docstring""" a__ : Dict = -sys.maxsize - 1 a__ : Tuple = n[:13] a__ : List[Any] = 13 while cur_index < len(_lowercase) - 13: if int(n[cur_index]) >= int(substr[0]): a__ : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: a__ : str = max(_lowercase , str_eval(_lowercase)) a__ : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'{solution() = }')
136
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
0
def UpperCAmelCase_ (_lowerCAmelCase : Any = 50 ): __UpperCamelCase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F"""{solution() = }""")
327
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
0
'''simple docstring''' def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Optional[int] = abs(lowerCamelCase__ ) A_ : Dict = 0 while n > 0: res += n % 10 n //= 10 return res def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Union[str, Any] = abs(lowerCamelCase__ ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def a ( lowerCamelCase__ ): '''simple docstring''' return sum(int(lowerCamelCase__ ) for c in str(abs(lowerCamelCase__ ) ) ) def a ( ): '''simple docstring''' from collections.abc import Callable from timeit import timeit def benchmark_a_function(lowerCamelCase__ , lowerCamelCase__ ) -> None: A_ : Tuple = f'{func.__name__}({value})' A_ : str = timeit(f'__main__.{call}' , setup="""import __main__""" ) print(f'{call:56} = {func(lowerCamelCase__ )} -- {timing:.4f} seconds' ) for value in (26_21_44, 11_25_89_99_06_84_26_24, 1_26_76_50_60_02_28_22_94_01_49_67_03_20_53_76): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(lowerCamelCase__ , lowerCamelCase__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
667
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> int: '''simple docstring''' assert column_title.isupper() lowerCAmelCase : Optional[Any] = 0 lowerCAmelCase : Optional[Any] = len(_UpperCAmelCase ) - 1 lowerCAmelCase : Optional[int] = 0 while index >= 0: lowerCAmelCase : Union[str, Any] = (ord(column_title[index] ) - 64) * pow(26, _UpperCAmelCase ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
343
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
0
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def a ( a , a , a ) ->str: '''simple docstring''' if isinstance(a , torch.Tensor ): return image elif isinstance(a , PIL.Image.Image ): SCREAMING_SNAKE_CASE = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE = np.concatenate(a , axis=0 ) SCREAMING_SNAKE_CASE = np.array(a ).astype(np.floataa ) / 2_55.0 SCREAMING_SNAKE_CASE = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE = torch.from_numpy(a ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE = torch.cat(a , dim=0 ) return image def a ( a , a , a , a=0.99_95 ) ->str: '''simple docstring''' if not isinstance(a , np.ndarray ): SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = va.device SCREAMING_SNAKE_CASE = va.cpu().numpy() SCREAMING_SNAKE_CASE = va.cpu().numpy() SCREAMING_SNAKE_CASE = np.sum(va * va / (np.linalg.norm(a ) * np.linalg.norm(a )) ) if np.abs(a ) > DOT_THRESHOLD: SCREAMING_SNAKE_CASE = (1 - t) * va + t * va else: SCREAMING_SNAKE_CASE = np.arccos(a ) SCREAMING_SNAKE_CASE = np.sin(a ) SCREAMING_SNAKE_CASE = theta_a * t SCREAMING_SNAKE_CASE = np.sin(a ) SCREAMING_SNAKE_CASE = np.sin(theta_a - theta_t ) / sin_theta_a SCREAMING_SNAKE_CASE = sin_theta_t / sin_theta_a SCREAMING_SNAKE_CASE = sa * va + sa * va if inputs_are_torch: SCREAMING_SNAKE_CASE = torch.from_numpy(a ).to(a ) return va def a ( a , a ) ->Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = F.normalize(a , dim=-1 ) SCREAMING_SNAKE_CASE = F.normalize(a , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def a ( a , a ) ->str: '''simple docstring''' for param in model.parameters(): SCREAMING_SNAKE_CASE = value class lowerCamelCase ( __snake_case ): def __init__( self :Optional[int] , lowercase :str , lowercase :int , lowercase :List[Any] , lowercase :Optional[Any] , lowercase :str , lowercase :str , lowercase :str , lowercase :List[Any]=None , lowercase :str=None , lowercase :Union[str, Any]=None , ) -> List[Any]: """simple docstring""" super().__init__() self.register_modules( vae=_lowerCAmelCase , text_encoder=_lowerCAmelCase , clip_model=_lowerCAmelCase , tokenizer=_lowerCAmelCase , unet=_lowerCAmelCase , scheduler=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , coca_model=_lowerCAmelCase , coca_tokenizer=_lowerCAmelCase , coca_transform=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE = ( feature_extractor.size if isinstance(feature_extractor.size , _lowerCAmelCase ) else feature_extractor.size['shortest_edge'] ) SCREAMING_SNAKE_CASE = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , _lowerCAmelCase ) set_requires_grad(self.clip_model , _lowerCAmelCase ) def snake_case__ ( self :int , lowercase :List[Any] = "auto" ) -> List[str]: """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory SCREAMING_SNAKE_CASE = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_lowerCAmelCase ) def snake_case__ ( self :Union[str, Any] ) -> Any: """simple docstring""" self.enable_attention_slicing(_lowerCAmelCase ) def snake_case__ ( self :Optional[Any] ) -> str: """simple docstring""" set_requires_grad(self.vae , _lowerCAmelCase ) def snake_case__ ( self :List[str] ) -> List[Any]: """simple docstring""" set_requires_grad(self.vae , _lowerCAmelCase ) def snake_case__ ( self :Dict ) -> List[str]: """simple docstring""" set_requires_grad(self.unet , _lowerCAmelCase ) def snake_case__ ( self :Tuple ) -> List[str]: """simple docstring""" set_requires_grad(self.unet , _lowerCAmelCase ) def snake_case__ ( self :Any , lowercase :Optional[int] , lowercase :Union[str, Any] , lowercase :Optional[Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE = min(int(num_inference_steps * strength ) , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = max(num_inference_steps - init_timestep , 0 ) SCREAMING_SNAKE_CASE = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def snake_case__ ( self :Optional[int] , lowercase :str , lowercase :Optional[int] , lowercase :List[Any] , lowercase :Union[str, Any] , lowercase :int , lowercase :int=None ) -> Any: """simple docstring""" if not isinstance(_lowerCAmelCase , torch.Tensor ): raise ValueError(f"""`image` has to be of type `torch.Tensor` but is {type(_lowerCAmelCase )}""" ) SCREAMING_SNAKE_CASE = image.to(device=_lowerCAmelCase , dtype=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): SCREAMING_SNAKE_CASE = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_lowerCAmelCase ) ] SCREAMING_SNAKE_CASE = torch.cat(_lowerCAmelCase , dim=0 ) else: SCREAMING_SNAKE_CASE = self.vae.encode(_lowerCAmelCase ).latent_dist.sample(_lowerCAmelCase ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE = 0.1_82_15 * init_latents SCREAMING_SNAKE_CASE = init_latents.repeat_interleave(_lowerCAmelCase , dim=0 ) SCREAMING_SNAKE_CASE = randn_tensor(init_latents.shape , generator=_lowerCAmelCase , device=_lowerCAmelCase , dtype=_lowerCAmelCase ) # get latents SCREAMING_SNAKE_CASE = self.scheduler.add_noise(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = init_latents return latents def snake_case__ ( self :List[Any] , lowercase :Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = self.coca_transform(_lowerCAmelCase ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): SCREAMING_SNAKE_CASE = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) SCREAMING_SNAKE_CASE = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split('''<end_of_text>''' )[0].replace('''<start_of_text>''' , '''''' ).rstrip(''' .,''' ) def snake_case__ ( self :Optional[Any] , lowercase :Tuple , lowercase :int ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE = self.feature_extractor.preprocess(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = torch.from_numpy(clip_image_input['''pixel_values'''][0] ).unsqueeze(0 ).to(self.device ).half() SCREAMING_SNAKE_CASE = self.clip_model.get_image_features(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = image_embeddings_clip.repeat_interleave(_lowerCAmelCase , dim=0 ) return image_embeddings_clip @torch.enable_grad() def snake_case__ ( self :Optional[int] , lowercase :Optional[Any] , lowercase :List[str] , lowercase :Optional[int] , lowercase :int , lowercase :Any , lowercase :Optional[int] , lowercase :List[str] , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE = latents.detach().requires_grad_() SCREAMING_SNAKE_CASE = self.scheduler.scale_model_input(_lowerCAmelCase , _lowerCAmelCase ) # predict the noise residual SCREAMING_SNAKE_CASE = self.unet(_lowerCAmelCase , _lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): SCREAMING_SNAKE_CASE = self.scheduler.alphas_cumprod[timestep] SCREAMING_SNAKE_CASE = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf SCREAMING_SNAKE_CASE = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 SCREAMING_SNAKE_CASE = torch.sqrt(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , _lowerCAmelCase ): SCREAMING_SNAKE_CASE = self.scheduler.sigmas[index] SCREAMING_SNAKE_CASE = latents - sigma * noise_pred else: raise ValueError(f"""scheduler type {type(self.scheduler )} not supported""" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE = 1 / 0.1_82_15 * sample SCREAMING_SNAKE_CASE = self.vae.decode(_lowerCAmelCase ).sample SCREAMING_SNAKE_CASE = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE = transforms.Resize(self.feature_extractor_size )(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = self.normalize(_lowerCAmelCase ).to(latents.dtype ) SCREAMING_SNAKE_CASE = self.clip_model.get_image_features(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = spherical_dist_loss(_lowerCAmelCase , _lowerCAmelCase ).mean() * clip_guidance_scale SCREAMING_SNAKE_CASE = -torch.autograd.grad(_lowerCAmelCase , _lowerCAmelCase )[0] if isinstance(self.scheduler , _lowerCAmelCase ): SCREAMING_SNAKE_CASE = latents.detach() + grads * (sigma**2) SCREAMING_SNAKE_CASE = noise_pred_original else: SCREAMING_SNAKE_CASE = noise_pred_original - torch.sqrt(_lowerCAmelCase ) * grads return noise_pred, latents @torch.no_grad() def __call__( self :List[Any] , lowercase :Optional[int] , lowercase :Tuple , lowercase :Any = None , lowercase :List[str] = None , lowercase :int = 5_1_2 , lowercase :List[str] = 5_1_2 , lowercase :Optional[int] = 0.6 , lowercase :List[str] = 5_0 , lowercase :Union[str, Any] = 7.5 , lowercase :str = 1 , lowercase :int = 0.0 , lowercase :Tuple = 1_0_0 , lowercase :List[Any] = None , lowercase :str = "pil" , lowercase :Optional[Any] = True , lowercase :List[str] = 0.8 , lowercase :Tuple = 0.1 , lowercase :Union[str, Any] = 0.1 , ) -> Tuple: """simple docstring""" if isinstance(_lowerCAmelCase , _lowerCAmelCase ) and len(_lowerCAmelCase ) != batch_size: raise ValueError(f"""You have passed {batch_size} batch_size, but only {len(_lowerCAmelCase )} generators.""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if isinstance(_lowerCAmelCase , torch.Generator ) and batch_size > 1: SCREAMING_SNAKE_CASE = [generator] + [None] * (batch_size - 1) SCREAMING_SNAKE_CASE = [ ('model', self.coca_model is None), ('tokenizer', self.coca_tokenizer is None), ('transform', self.coca_transform is None), ] SCREAMING_SNAKE_CASE = [x[0] for x in coca_is_none if x[1]] SCREAMING_SNAKE_CASE = ', '.join(_lowerCAmelCase ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(_lowerCAmelCase ): raise ValueError( f"""Content prompt is None and CoCa [{coca_is_none_str}] is None.""" f"""Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) SCREAMING_SNAKE_CASE = self.get_image_description(_lowerCAmelCase ) if style_prompt is None: if len(_lowerCAmelCase ): raise ValueError( f"""Style prompt is None and CoCa [{coca_is_none_str}] is None.""" f""" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) SCREAMING_SNAKE_CASE = self.get_image_description(_lowerCAmelCase ) # get prompt text embeddings for content and style SCREAMING_SNAKE_CASE = self.tokenizer( _lowerCAmelCase , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=_lowerCAmelCase , return_tensors='''pt''' , ) SCREAMING_SNAKE_CASE = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] SCREAMING_SNAKE_CASE = self.tokenizer( _lowerCAmelCase , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=_lowerCAmelCase , return_tensors='''pt''' , ) SCREAMING_SNAKE_CASE = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] SCREAMING_SNAKE_CASE = slerp(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # duplicate text embeddings for each generation per prompt SCREAMING_SNAKE_CASE = text_embeddings.repeat_interleave(_lowerCAmelCase , dim=0 ) # set timesteps SCREAMING_SNAKE_CASE = 'offset' in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) SCREAMING_SNAKE_CASE = {} if accepts_offset: SCREAMING_SNAKE_CASE = 1 self.scheduler.set_timesteps(_lowerCAmelCase , **_lowerCAmelCase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) SCREAMING_SNAKE_CASE = self.get_timesteps(_lowerCAmelCase , _lowerCAmelCase , self.device ) SCREAMING_SNAKE_CASE = timesteps[:1].repeat(_lowerCAmelCase ) # Preprocess image SCREAMING_SNAKE_CASE = preprocess(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = self.prepare_latents( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , text_embeddings.dtype , self.device , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = preprocess(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = self.prepare_latents( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , text_embeddings.dtype , self.device , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = slerp(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if clip_guidance_scale > 0: SCREAMING_SNAKE_CASE = self.get_clip_image_embeddings(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = self.get_clip_image_embeddings(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = slerp( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. SCREAMING_SNAKE_CASE = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: SCREAMING_SNAKE_CASE = content_text_input.input_ids.shape[-1] SCREAMING_SNAKE_CASE = self.tokenizer([''''''] , padding='''max_length''' , max_length=_lowerCAmelCase , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt SCREAMING_SNAKE_CASE = uncond_embeddings.repeat_interleave(_lowerCAmelCase , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes SCREAMING_SNAKE_CASE = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. SCREAMING_SNAKE_CASE = (batch_size, self.unet.config.in_channels, height // 8, width // 8) SCREAMING_SNAKE_CASE = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps SCREAMING_SNAKE_CASE = torch.randn(_lowerCAmelCase , generator=_lowerCAmelCase , device='''cpu''' , dtype=_lowerCAmelCase ).to( self.device ) else: SCREAMING_SNAKE_CASE = torch.randn(_lowerCAmelCase , generator=_lowerCAmelCase , device=self.device , dtype=_lowerCAmelCase ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) SCREAMING_SNAKE_CASE = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler SCREAMING_SNAKE_CASE = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] SCREAMING_SNAKE_CASE = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) SCREAMING_SNAKE_CASE = {} if accepts_eta: SCREAMING_SNAKE_CASE = eta # check if the scheduler accepts generator SCREAMING_SNAKE_CASE = 'generator' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: SCREAMING_SNAKE_CASE = generator with self.progress_bar(total=_lowerCAmelCase ): for i, t in enumerate(_lowerCAmelCase ): # expand the latents if we are doing classifier free guidance SCREAMING_SNAKE_CASE = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents SCREAMING_SNAKE_CASE = self.scheduler.scale_model_input(_lowerCAmelCase , _lowerCAmelCase ) # predict the noise residual SCREAMING_SNAKE_CASE = self.unet(_lowerCAmelCase , _lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase ).sample # perform classifier free guidance if do_classifier_free_guidance: SCREAMING_SNAKE_CASE = noise_pred.chunk(2 ) SCREAMING_SNAKE_CASE = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: SCREAMING_SNAKE_CASE = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) SCREAMING_SNAKE_CASE = self.cond_fn( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) # compute the previous noisy sample x_t -> x_t-1 SCREAMING_SNAKE_CASE = self.scheduler.step(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , **_lowerCAmelCase ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE = 1 / 0.1_82_15 * latents SCREAMING_SNAKE_CASE = self.vae.decode(_lowerCAmelCase ).sample SCREAMING_SNAKE_CASE = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE = self.numpy_to_pil(_lowerCAmelCase ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=_lowerCAmelCase , nsfw_content_detected=_lowerCAmelCase )
201
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
0
'''simple docstring''' from functools import lru_cache def UpperCAmelCase_ ( __lowercase : List[str] ) -> set: '''simple docstring''' _UpperCAmelCase = 2 _UpperCAmelCase = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(__lowercase ) if n > 1: factors.add(__lowercase ) return factors @lru_cache def UpperCAmelCase_ ( __lowercase : Union[str, Any] ) -> int: '''simple docstring''' return len(unique_prime_factors(__lowercase ) ) def UpperCAmelCase_ ( __lowercase : Dict ) -> bool: '''simple docstring''' return len(set(__lowercase ) ) in (0, 1) def UpperCAmelCase_ ( __lowercase : int ) -> list: '''simple docstring''' _UpperCAmelCase = 2 while True: # Increment each value of a generated range _UpperCAmelCase = [base + i for i in range(__lowercase )] # Run elements through out unique_prime_factors function # Append our target number to the end. _UpperCAmelCase = [upf_len(__lowercase ) for x in group] checker.append(__lowercase ) # If all numbers in the list are equal, return the group variable. if equality(__lowercase ): return group # Increment our base variable by 1 base += 1 def UpperCAmelCase_ ( __lowercase : int = 4 ) -> int: '''simple docstring''' _UpperCAmelCase = run(__lowercase ) return results[0] if len(__lowercase ) else None if __name__ == "__main__": print(solution())
236
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
0
import random def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = False ): lowerCamelCase_ = {i: [] for i in range(lowerCamelCase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowerCamelCase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowerCamelCase__ ): for j in range(i + 1 , lowerCamelCase__ ): if random.random() < probability: graph[i].append(lowerCamelCase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowerCamelCase__ ) return graph def lowerCamelCase_ ( lowerCamelCase__ ): return { i: [j for j in range(lowerCamelCase__ ) if i != j] for i in range(lowerCamelCase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
463
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
"""simple docstring""" import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def A_ ( snake_case_ : str = 8 ): '''simple docstring''' UpperCamelCase : Any = ascii_letters + digits + punctuation return "".join(secrets.choice(snake_case_ ) for _ in range(snake_case_ ) ) def A_ ( snake_case_ : Union[str, Any] ,snake_case_ : str ): '''simple docstring''' # Password Generator = full boot with random_number, random_letters, and # random_character FUNCTIONS # Put your code here... i -= len(snake_case_ ) UpperCamelCase : Any = i // 3 UpperCamelCase : Union[str, Any] = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) UpperCamelCase : Dict = ( chars_incl + random(snake_case_ ,quotient + remainder ) + random(snake_case_ ,snake_case_ ) + random(snake_case_ ,snake_case_ ) ) UpperCamelCase : Any = list(snake_case_ ) shuffle(snake_case_ ) return "".join(snake_case_ ) # random is a generalised function for letters, characters and numbers def A_ ( snake_case_ : Any ,snake_case_ : Any ): '''simple docstring''' return "".join(secrets.choice(snake_case_ ) for _ in range(snake_case_ ) ) def A_ ( snake_case_ : List[Any] ,snake_case_ : Optional[Any] ): '''simple docstring''' pass # Put your code here... def A_ ( snake_case_ : Any ,snake_case_ : int ): '''simple docstring''' pass # Put your code here... def A_ ( snake_case_ : Any ,snake_case_ : int ): '''simple docstring''' pass # Put your code here... def A_ ( snake_case_ : int ,snake_case_ : List[Any] = 8 ): '''simple docstring''' if len(snake_case_ ) < min_length: # Your Password must be at least 8 characters long return False UpperCamelCase : List[str] = any(char in ascii_uppercase for char in password ) UpperCamelCase : Union[str, Any] = any(char in ascii_lowercase for char in password ) UpperCamelCase : Optional[int] = any(char in digits for char in password ) UpperCamelCase : int = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def A_ ( ): '''simple docstring''' UpperCamelCase : Tuple = int(input("""Please indicate the max length of your password: """ ).strip() ) UpperCamelCase : Optional[Any] = input( """Please indicate the characters that must be in your password: """ ).strip() print("""Password generated:""" ,password_generator(snake_case_ ) ) print( """Alternative Password generated:""" ,alternative_password_generator(snake_case_ ,snake_case_ ) ,) print("""[If you are thinking of using this passsword, You better save it.]""" ) if __name__ == "__main__": main()
499
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
0
'''simple docstring''' import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) lowercase = { '''sample_size''': 32, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': 1000, '''block_out_channels''': [32, 64], '''attention_head_dim''': 8, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowercase = { '''sample_size''': 64, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 3, '''num_class_embeds''': 1000, '''block_out_channels''': [192, 192 * 2, 192 * 3, 192 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowercase = { '''sample_size''': 256, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': None, '''block_out_channels''': [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''default''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowercase = { '''num_train_timesteps''': 40, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowercase = { '''num_train_timesteps''': 201, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowercase = { '''num_train_timesteps''': 151, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } def __A ( _SCREAMING_SNAKE_CASE : int ): """simple docstring""" if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("boolean value expected" ) def __A ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any]=False ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = checkpoint[f'{old_prefix}.in_layers.0.weight'] __SCREAMING_SNAKE_CASE : int = checkpoint[f'{old_prefix}.in_layers.0.bias'] __SCREAMING_SNAKE_CASE : Optional[int] = checkpoint[f'{old_prefix}.in_layers.2.weight'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint[f'{old_prefix}.in_layers.2.bias'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint[f'{old_prefix}.emb_layers.1.weight'] __SCREAMING_SNAKE_CASE : Optional[int] = checkpoint[f'{old_prefix}.emb_layers.1.bias'] __SCREAMING_SNAKE_CASE : Union[str, Any] = checkpoint[f'{old_prefix}.out_layers.0.weight'] __SCREAMING_SNAKE_CASE : List[str] = checkpoint[f'{old_prefix}.out_layers.0.bias'] __SCREAMING_SNAKE_CASE : Optional[int] = checkpoint[f'{old_prefix}.out_layers.3.weight'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint[f'{old_prefix}.out_layers.3.bias'] if has_skip: __SCREAMING_SNAKE_CASE : Union[str, Any] = checkpoint[f'{old_prefix}.skip_connection.weight'] __SCREAMING_SNAKE_CASE : int = checkpoint[f'{old_prefix}.skip_connection.bias'] return new_checkpoint def __A ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Optional[Any]=None ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = checkpoint[f'{old_prefix}.qkv.weight'].chunk(3 , dim=0 ) __SCREAMING_SNAKE_CASE : Union[str, Any] = checkpoint[f'{old_prefix}.qkv.bias'].chunk(3 , dim=0 ) __SCREAMING_SNAKE_CASE : str = checkpoint[f'{old_prefix}.norm.weight'] __SCREAMING_SNAKE_CASE : Tuple = checkpoint[f'{old_prefix}.norm.bias'] __SCREAMING_SNAKE_CASE : int = weight_q.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : List[Any] = bias_q.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : int = weight_k.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : Any = bias_k.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : List[str] = weight_v.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : List[str] = bias_v.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : Tuple = ( checkpoint[f'{old_prefix}.proj_out.weight'].squeeze(-1 ).squeeze(-1 ) ) __SCREAMING_SNAKE_CASE : Dict = checkpoint[f'{old_prefix}.proj_out.bias'].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def __A ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = torch.load(_SCREAMING_SNAKE_CASE , map_location="cpu" ) __SCREAMING_SNAKE_CASE : str = {} __SCREAMING_SNAKE_CASE : str = checkpoint['time_embed.0.weight'] __SCREAMING_SNAKE_CASE : Tuple = checkpoint['time_embed.0.bias'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint['time_embed.2.weight'] __SCREAMING_SNAKE_CASE : int = checkpoint['time_embed.2.bias'] if unet_config["num_class_embeds"] is not None: __SCREAMING_SNAKE_CASE : Dict = checkpoint['label_emb.weight'] __SCREAMING_SNAKE_CASE : Tuple = checkpoint['input_blocks.0.0.weight'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint['input_blocks.0.0.bias'] __SCREAMING_SNAKE_CASE : Union[str, Any] = unet_config['down_block_types'] __SCREAMING_SNAKE_CASE : Tuple = unet_config['layers_per_block'] __SCREAMING_SNAKE_CASE : Any = unet_config['attention_head_dim'] __SCREAMING_SNAKE_CASE : List[Any] = unet_config['block_out_channels'] __SCREAMING_SNAKE_CASE : Optional[Any] = 1 __SCREAMING_SNAKE_CASE : List[str] = channels_list[0] for i, layer_type in enumerate(_SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Tuple = channels_list[i] __SCREAMING_SNAKE_CASE : Tuple = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(_SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : str = f'down_blocks.{i}.resnets.{j}' __SCREAMING_SNAKE_CASE : Dict = f'input_blocks.{current_layer}.0' __SCREAMING_SNAKE_CASE : int = True if j == 0 and downsample_block_has_skip else False __SCREAMING_SNAKE_CASE : Optional[int] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , has_skip=_SCREAMING_SNAKE_CASE ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(_SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Tuple = f'down_blocks.{i}.resnets.{j}' __SCREAMING_SNAKE_CASE : Union[str, Any] = f'input_blocks.{current_layer}.0' __SCREAMING_SNAKE_CASE : List[Any] = True if j == 0 and downsample_block_has_skip else False __SCREAMING_SNAKE_CASE : Union[str, Any] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , has_skip=_SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Dict = f'down_blocks.{i}.attentions.{j}' __SCREAMING_SNAKE_CASE : int = f'input_blocks.{current_layer}.1' __SCREAMING_SNAKE_CASE : Any = convert_attention( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(_SCREAMING_SNAKE_CASE ) - 1: __SCREAMING_SNAKE_CASE : int = f'down_blocks.{i}.downsamplers.0' __SCREAMING_SNAKE_CASE : List[Any] = f'input_blocks.{current_layer}.0' __SCREAMING_SNAKE_CASE : Tuple = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) current_layer += 1 __SCREAMING_SNAKE_CASE : List[Any] = current_channels # hardcoded the mid-block for now __SCREAMING_SNAKE_CASE : List[Any] = 'mid_block.resnets.0' __SCREAMING_SNAKE_CASE : Union[str, Any] = 'middle_block.0' __SCREAMING_SNAKE_CASE : Optional[int] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Tuple = 'mid_block.attentions.0' __SCREAMING_SNAKE_CASE : Any = 'middle_block.1' __SCREAMING_SNAKE_CASE : str = convert_attention(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Optional[Any] = 'mid_block.resnets.1' __SCREAMING_SNAKE_CASE : Optional[Any] = 'middle_block.2' __SCREAMING_SNAKE_CASE : List[Any] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Optional[int] = 0 __SCREAMING_SNAKE_CASE : List[str] = unet_config['up_block_types'] for i, layer_type in enumerate(_SCREAMING_SNAKE_CASE ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): __SCREAMING_SNAKE_CASE : Tuple = f'up_blocks.{i}.resnets.{j}' __SCREAMING_SNAKE_CASE : int = f'output_blocks.{current_layer}.0' __SCREAMING_SNAKE_CASE : str = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , has_skip=_SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(_SCREAMING_SNAKE_CASE ) - 1: __SCREAMING_SNAKE_CASE : Dict = f'up_blocks.{i}.upsamplers.0' __SCREAMING_SNAKE_CASE : Optional[Any] = f'output_blocks.{current_layer-1}.1' __SCREAMING_SNAKE_CASE : Dict = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): __SCREAMING_SNAKE_CASE : Tuple = f'up_blocks.{i}.resnets.{j}' __SCREAMING_SNAKE_CASE : List[str] = f'output_blocks.{current_layer}.0' __SCREAMING_SNAKE_CASE : Optional[Any] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , has_skip=_SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : str = f'up_blocks.{i}.attentions.{j}' __SCREAMING_SNAKE_CASE : Dict = f'output_blocks.{current_layer}.1' __SCREAMING_SNAKE_CASE : Optional[Any] = convert_attention( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(_SCREAMING_SNAKE_CASE ) - 1: __SCREAMING_SNAKE_CASE : int = f'up_blocks.{i}.upsamplers.0' __SCREAMING_SNAKE_CASE : str = f'output_blocks.{current_layer-1}.2' __SCREAMING_SNAKE_CASE : List[Any] = convert_resnet(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : List[Any] = checkpoint['out.0.weight'] __SCREAMING_SNAKE_CASE : Optional[Any] = checkpoint['out.0.bias'] __SCREAMING_SNAKE_CASE : Dict = checkpoint['out.2.weight'] __SCREAMING_SNAKE_CASE : Dict = checkpoint['out.2.bias'] return new_checkpoint if __name__ == "__main__": lowercase = argparse.ArgumentParser() parser.add_argument('''--unet_path''', default=None, type=str, required=True, help='''Path to the unet.pt to convert.''') parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output the converted UNet model.''' ) parser.add_argument('''--class_cond''', default=True, type=str, help='''Whether the model is class-conditional.''') lowercase = parser.parse_args() lowercase = strabool(args.class_cond) lowercase = os.path.basename(args.unet_path) print(F"""Checkpoint: {ckpt_name}""") # Get U-Net config if "imagenet64" in ckpt_name: lowercase = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowercase = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: lowercase = TEST_UNET_CONFIG else: raise ValueError(F"""Checkpoint type {ckpt_name} is not currently supported.""") if not args.class_cond: lowercase = None lowercase = con_pt_to_diffuser(args.unet_path, unet_config) lowercase = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: lowercase = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: lowercase = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowercase = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F"""Checkpoint type {ckpt_name} is not currently supported.""") lowercase = CMStochasticIterativeScheduler(**scheduler_config) lowercase = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
211
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
0
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/config.json''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/config.json''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/config.json''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/config.json''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/config.json''', } class __magic_name__ ( __snake_case ): UpperCamelCase__ = "t5" UpperCamelCase__ = ["past_key_values"] UpperCamelCase__ = {"hidden_size": "d_model", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self , snake_case_=3_21_28 , snake_case_=5_12 , snake_case_=64 , snake_case_=20_48 , snake_case_=6 , snake_case_=None , snake_case_=8 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="relu" , snake_case_=True , snake_case_=True , snake_case_=0 , snake_case_=1 , **snake_case_ , ): lowercase =vocab_size lowercase =d_model lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =feed_forward_proj lowercase =use_cache lowercase =self.feed_forward_proj.split('''-''' ) lowercase =act_info[-1] lowercase =act_info[0] == 'gated' if len(_lowerCAmelCase ) > 1 and act_info[0] != "gated" or len(_lowerCAmelCase ) > 2: raise ValueError( f'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowercase ='gelu_new' super().__init__( pad_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase , ) class __magic_name__ ( __snake_case ): @property def _A( self ): lowercase ={ 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: lowercase ='past_encoder_sequence + sequence' lowercase ={0: 'batch'} lowercase ={0: 'batch', 1: 'past_decoder_sequence + sequence'} else: lowercase ={0: 'batch', 1: 'decoder_sequence'} lowercase ={0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(_lowerCAmelCase , direction='''inputs''' ) return common_inputs @property def _A( self ): return 13
72
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
0
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class A ( __snake_case , unittest.TestCase ): __UpperCAmelCase : Dict = XLNetTokenizer __UpperCAmelCase : str = XLNetTokenizerFast __UpperCAmelCase : Optional[Any] = True __UpperCAmelCase : int = True def lowercase_ (self : Optional[int] ) -> int: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ = XLNetTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ (self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = '<s>' UpperCAmelCase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def lowercase_ (self : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "<eod>" ) self.assertEqual(len(_lowerCAmelCase ) , 1_0_0_6 ) def lowercase_ (self : str ) -> Any: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = XLNetTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.tokenize("This is a test" ) self.assertListEqual(_lowerCAmelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] ) UpperCAmelCase__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) UpperCAmelCase__ = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] ) UpperCAmelCase__ = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def lowercase_ (self : List[Any] ) -> int: """simple docstring""" UpperCAmelCase__ = XLNetTokenizer(_lowerCAmelCase , do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + "", "i", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["▁he", "ll", "o"] ) def lowercase_ (self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase__ = XLNetTokenizer(_lowerCAmelCase , do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) @slow def lowercase_ (self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = XLNetTokenizer.from_pretrained("xlnet-base-cased" ) UpperCAmelCase__ = tokenizer.encode("sequence builders" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.encode("multi-sequence build" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def lowercase_ (self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = {'input_ids': [[1_7, 2_1_4_4_2, 2_7_0, 1_7, 1_0, 1_4_6_4_5, 3_1_8, 3_4, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 7_7_5_2, 2_2_0_1_8, 2_3, 2_1, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 3_3_5_2, 1_4_4_3_1, 1_3, 5_5_0_0, 1_1, 1_1_7_6, 5_8_0, 1_3, 1_6_8_1_9, 4_7_9_7, 2_3, 1_7, 1_0, 1_7_1_3_5, 6_5_8, 1_9, 4_5_7, 7_9_3_2, 1_3, 1_8_4, 1_9, 3_1_5_4, 1_7_1_3_5, 6_4_6_8, 1_9, 1_4_0_4, 1_2_2_6_9, 1_9, 4_2_2_9, 5_3_5_6, 1_6_2_6_4, 4_6, 1_9, 1_7, 2_0_5_4_5, 1_0_3_9_5, 9, 9, 9, 1_1, 2_8, 6_4_2_1, 9_5_3_1, 2_0_7_2_9, 1_7, 1_0, 3_5_3, 1_7_0_2_2, 1_1, 2_1, 6_4_2_1, 9_5_3_1, 1_6_9_4_9, 1_7, 1_0, 1_1_5_0_9, 7_5_3, 1_1, 3_3, 9_5, 2_4_2_1, 7_3_8_5, 9_5_6, 1_4_4_3_1, 2_6_2_6, 2_5, 8_4_2, 7_3_8_5, 4_8_3_6, 2_1, 1_4_2_9, 2_2_7_2, 9_8_5_5, 3_1_2_0, 1_6_1, 2_4_7_3_8, 1_9, 1_3_2_0_3, 6_5_8, 2_1_8, 7_8_7, 2_1, 4_3_0, 1_8_4_8_2, 8_4_7, 2_6_3_7, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2_2, 2_2_1_7_8, 2_7, 1_0_6_4, 2_2, 9_5_6, 1_3, 1_1_1_0_1, 1_4_2_9, 5_8_5_4, 2_4_3_1_3, 1_8_9_5_3, 4_0, 4_2_2, 2_4_3_6_6, 6_8, 1_7_5_8, 3_7, 1_0_4_8_3, 1_4_2_5_7, 3_1, 2_0_7, 2_6_3, 2_1, 2_0_3, 3_7_7_3, 2_5, 7_1, 9_7_3_5, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2, 2_0_4_9, 3_4_4_2, 1_7, 1_3_8_9_4, 3_3_8_0, 2_3, 9_5, 1_8, 1_7_6_3_4, 2_2_8_8, 9, 4, 3]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name="xlnet-base-cased" , revision="c841166438c31ec7ca9a106dee7bb312b73ae511" , )
486
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase : Tuple ={ "configuration_squeezebert": [ "SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "SqueezeBertConfig", "SqueezeBertOnnxConfig", ], "tokenization_squeezebert": ["SqueezeBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[int] =["SqueezeBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : int =[ "SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "SqueezeBertForMaskedLM", "SqueezeBertForMultipleChoice", "SqueezeBertForQuestionAnswering", "SqueezeBertForSequenceClassification", "SqueezeBertForTokenClassification", "SqueezeBertModel", "SqueezeBertModule", "SqueezeBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _lowercase : Optional[int] =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
136
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
def UpperCAmelCase_ (_lowerCAmelCase : List[Any] ): if len(_lowerCAmelCase ) <= 1: return [tuple(_lowerCAmelCase )] __UpperCamelCase : List[Any] = [] def generate(_lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): __UpperCamelCase : Tuple = [0] * n res.append(tuple(_lowerCAmelCase ) ) __UpperCamelCase : int = 0 while i < n: if c[i] < i: if i % 2 == 0: __UpperCamelCase : Optional[Any] = arr[i], arr[0] else: __UpperCamelCase : Union[str, Any] = arr[i], arr[c[i]] res.append(tuple(_lowerCAmelCase ) ) c[i] += 1 __UpperCamelCase : Tuple = 0 else: __UpperCamelCase : str = 0 i += 1 generate(len(_lowerCAmelCase ) , _lowerCAmelCase ) return res if __name__ == "__main__": lowercase : Optional[Any] = input("Enter numbers separated by a comma:\n").strip() lowercase : Dict = [int(item) for item in user_input.split(",")] print(heaps(arr))
327
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase :Any = { '''configuration_m2m_100''': ['''M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''M2M100Config''', '''M2M100OnnxConfig'''], '''tokenization_m2m_100''': ['''M2M100Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase :Tuple = [ '''M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST''', '''M2M100ForConditionalGeneration''', '''M2M100Model''', '''M2M100PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys lowerCamelCase :Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
667
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __A : int = { '''configuration_gpt_neo''': ['''GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoConfig''', '''GPTNeoOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ '''GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoForCausalLM''', '''GPTNeoForQuestionAnswering''', '''GPTNeoForSequenceClassification''', '''GPTNeoForTokenClassification''', '''GPTNeoModel''', '''GPTNeoPreTrainedModel''', '''load_tf_weights_in_gpt_neo''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = [ '''FlaxGPTNeoForCausalLM''', '''FlaxGPTNeoModel''', '''FlaxGPTNeoPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
343
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase = logging.get_logger(__name__) def a ( a ) ->YolosConfig: '''simple docstring''' SCREAMING_SNAKE_CASE = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: SCREAMING_SNAKE_CASE = 192 SCREAMING_SNAKE_CASE = 768 SCREAMING_SNAKE_CASE = 12 SCREAMING_SNAKE_CASE = 3 SCREAMING_SNAKE_CASE = [800, 1333] SCREAMING_SNAKE_CASE = False elif yolos_name == "yolos_s_dWr": SCREAMING_SNAKE_CASE = 330 SCREAMING_SNAKE_CASE = 14 SCREAMING_SNAKE_CASE = 6 SCREAMING_SNAKE_CASE = 1320 elif "yolos_s" in yolos_name: SCREAMING_SNAKE_CASE = 384 SCREAMING_SNAKE_CASE = 1536 SCREAMING_SNAKE_CASE = 12 SCREAMING_SNAKE_CASE = 6 elif "yolos_b" in yolos_name: SCREAMING_SNAKE_CASE = [800, 1344] SCREAMING_SNAKE_CASE = 91 SCREAMING_SNAKE_CASE = 'huggingface/label-files' SCREAMING_SNAKE_CASE = 'coco-detection-id2label.json' SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) SCREAMING_SNAKE_CASE = {int(a ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = idalabel SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config def a ( a , a , a = False ) ->str: '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) SCREAMING_SNAKE_CASE = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE = in_proj_weight[: config.hidden_size, :] SCREAMING_SNAKE_CASE = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE = in_proj_weight[-config.hidden_size :, :] SCREAMING_SNAKE_CASE = in_proj_bias[-config.hidden_size :] def a ( a ) ->str: '''simple docstring''' if "backbone" in name: SCREAMING_SNAKE_CASE = name.replace('''backbone''' , '''vit''' ) if "cls_token" in name: SCREAMING_SNAKE_CASE = name.replace('''cls_token''' , '''embeddings.cls_token''' ) if "det_token" in name: SCREAMING_SNAKE_CASE = name.replace('''det_token''' , '''embeddings.detection_tokens''' ) if "mid_pos_embed" in name: SCREAMING_SNAKE_CASE = name.replace('''mid_pos_embed''' , '''encoder.mid_position_embeddings''' ) if "pos_embed" in name: SCREAMING_SNAKE_CASE = name.replace('''pos_embed''' , '''embeddings.position_embeddings''' ) if "patch_embed.proj" in name: SCREAMING_SNAKE_CASE = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "blocks" in name: SCREAMING_SNAKE_CASE = name.replace('''blocks''' , '''encoder.layer''' ) if "attn.proj" in name: SCREAMING_SNAKE_CASE = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: SCREAMING_SNAKE_CASE = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: SCREAMING_SNAKE_CASE = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: SCREAMING_SNAKE_CASE = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: SCREAMING_SNAKE_CASE = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: SCREAMING_SNAKE_CASE = name.replace('''mlp.fc2''' , '''output.dense''' ) if "class_embed" in name: SCREAMING_SNAKE_CASE = name.replace('''class_embed''' , '''class_labels_classifier''' ) if "bbox_embed" in name: SCREAMING_SNAKE_CASE = name.replace('''bbox_embed''' , '''bbox_predictor''' ) if "vit.norm" in name: SCREAMING_SNAKE_CASE = name.replace('''vit.norm''' , '''vit.layernorm''' ) return name def a ( a , a ) ->dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): SCREAMING_SNAKE_CASE = orig_state_dict.pop(a ) if "qkv" in key: SCREAMING_SNAKE_CASE = key.split('''.''' ) SCREAMING_SNAKE_CASE = int(key_split[2] ) SCREAMING_SNAKE_CASE = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: SCREAMING_SNAKE_CASE = val[:dim, :] SCREAMING_SNAKE_CASE = val[ dim : dim * 2, : ] SCREAMING_SNAKE_CASE = val[-dim:, :] else: SCREAMING_SNAKE_CASE = val[:dim] SCREAMING_SNAKE_CASE = val[dim : dim * 2] SCREAMING_SNAKE_CASE = val[-dim:] else: SCREAMING_SNAKE_CASE = val return orig_state_dict def a ( ) ->torch.Tensor: '''simple docstring''' SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def a ( a , a , a , a = False ) ->Any: '''simple docstring''' SCREAMING_SNAKE_CASE = get_yolos_config(a ) # load original state_dict SCREAMING_SNAKE_CASE = torch.load(a , map_location='''cpu''' )['model'] # load 🤗 model SCREAMING_SNAKE_CASE = YolosForObjectDetection(a ) model.eval() SCREAMING_SNAKE_CASE = convert_state_dict(a , a ) model.load_state_dict(a ) # Check outputs on an image, prepared by YolosImageProcessor SCREAMING_SNAKE_CASE = 800 if yolos_name != 'yolos_ti' else 512 SCREAMING_SNAKE_CASE = YolosImageProcessor(format='''coco_detection''' , size=a ) SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE = model(**a ) SCREAMING_SNAKE_CASE = outputs.logits, outputs.pred_boxes SCREAMING_SNAKE_CASE = None, None if yolos_name == "yolos_ti": SCREAMING_SNAKE_CASE = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) SCREAMING_SNAKE_CASE = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": SCREAMING_SNAKE_CASE = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) SCREAMING_SNAKE_CASE = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": SCREAMING_SNAKE_CASE = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) SCREAMING_SNAKE_CASE = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": SCREAMING_SNAKE_CASE = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) SCREAMING_SNAKE_CASE = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": SCREAMING_SNAKE_CASE = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) SCREAMING_SNAKE_CASE = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"""Unknown yolos_name: {yolos_name}""" ) assert torch.allclose(logits[0, :3, :3] , a , atol=1E-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , a , atol=1E-4 ) Path(a ).mkdir(exist_ok=a ) print(F"""Saving model {yolos_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if push_to_hub: SCREAMING_SNAKE_CASE = { 'yolos_ti': 'yolos-tiny', 'yolos_s_200_pre': 'yolos-small', 'yolos_s_300_pre': 'yolos-small-300', 'yolos_s_dWr': 'yolos-small-dwr', 'yolos_base': 'yolos-base', } print('''Pushing to the hub...''' ) SCREAMING_SNAKE_CASE = model_mapping[yolos_name] image_processor.push_to_hub(a , organization='''hustvl''' ) model.push_to_hub(a , organization='''hustvl''' ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--yolos_name', default='yolos_s_200_pre', type=str, help=( 'Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',' ' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.' ), ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original state dict (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __lowerCAmelCase = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
201
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
'''simple docstring''' import requests from bsa import BeautifulSoup def UpperCAmelCase_ ( __lowercase : int = "https://www.worldometers.info/coronavirus" ) -> dict: '''simple docstring''' _UpperCAmelCase = BeautifulSoup(requests.get(__lowercase ).text , "html.parser" ) _UpperCAmelCase = soup.findAll("h1" ) _UpperCAmelCase = soup.findAll("div" , {"class": "maincounter-number"} ) keys += soup.findAll("span" , {"class": "panel-title"} ) values += soup.findAll("div" , {"class": "number-table-main"} ) return {key.text.strip(): value.text.strip() for key, value in zip(__lowercase , __lowercase )} if __name__ == "__main__": print('''\033[1m''' + '''COVID-19 Status of the World''' + '''\033[0m\n''') for key, value in world_covidaa_stats().items(): print(F"{key}\n{value}\n")
236
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
0
import logging import os import threading import time try: import warnings except ImportError: __A =None try: import msvcrt except ImportError: __A =None try: import fcntl except ImportError: __A =None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __A =OSError # Data # ------------------------------------------------ __A =[ '''Timeout''', '''BaseFileLock''', '''WindowsFileLock''', '''UnixFileLock''', '''SoftFileLock''', '''FileLock''', ] __A ='''3.0.12''' __A =None def lowerCamelCase_ ( ): global _logger lowerCamelCase_ = _logger or logging.getLogger(__name__ ) return _logger class _SCREAMING_SNAKE_CASE ( __snake_case ): def __init__( self , lowercase ) -> Optional[int]: lowerCamelCase_ = lock_file return None def __str__( self ) -> List[Any]: lowerCamelCase_ = f'The file lock \'{self.lock_file}\' could not be acquired.' return temp class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase ) -> Union[str, Any]: lowerCamelCase_ = lock return None def __enter__( self ) -> List[Any]: return self.lock def __exit__( self , lowercase , lowercase , lowercase ) -> List[str]: self.lock.release() return None class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase , lowercase=-1 , lowercase=None ) -> List[Any]: lowerCamelCase_ = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long lowerCamelCase_ = self.hash_filename_if_too_long(_lowerCAmelCase , _lowerCAmelCase ) # The path to the lock file. lowerCamelCase_ = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. lowerCamelCase_ = None # The default timeout value. lowerCamelCase_ = timeout # We use this lock primarily for the lock counter. lowerCamelCase_ = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. lowerCamelCase_ = 0 return None @property def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: return self._lock_file @property def SCREAMING_SNAKE_CASE_( self ) -> Tuple: return self._timeout @timeout.setter def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Any: lowerCamelCase_ = float(_lowerCAmelCase ) return None def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: raise NotImplementedError() def SCREAMING_SNAKE_CASE_( self ) -> Dict: raise NotImplementedError() @property def SCREAMING_SNAKE_CASE_( self ) -> List[str]: return self._lock_file_fd is not None def SCREAMING_SNAKE_CASE_( self , lowercase=None , lowercase=0.0_5 ) -> Tuple: # Use the default timeout, if no timeout is provided. if timeout is None: lowerCamelCase_ = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 lowerCamelCase_ = id(self ) lowerCamelCase_ = self._lock_file lowerCamelCase_ = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(f'Attempting to acquire lock {lock_id} on {lock_filename}' ) self._acquire() if self.is_locked: logger().debug(f'Lock {lock_id} acquired on {lock_filename}' ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(f'Timeout on acquiring lock {lock_id} on {lock_filename}' ) raise Timeout(self._lock_file ) else: logger().debug( f'Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...' ) time.sleep(_lowerCAmelCase ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: lowerCamelCase_ = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def SCREAMING_SNAKE_CASE_( self , lowercase=False ) -> Dict: with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: lowerCamelCase_ = id(self ) lowerCamelCase_ = self._lock_file logger().debug(f'Attempting to release lock {lock_id} on {lock_filename}' ) self._release() lowerCamelCase_ = 0 logger().debug(f'Lock {lock_id} released on {lock_filename}' ) return None def __enter__( self ) -> Union[str, Any]: self.acquire() return self def __exit__( self , lowercase , lowercase , lowercase ) -> Optional[Any]: self.release() return None def __del__( self ) -> Optional[Any]: self.release(force=_lowerCAmelCase ) return None def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> Optional[int]: lowerCamelCase_ = os.path.basename(_lowerCAmelCase ) if len(_lowerCAmelCase ) > max_length and max_length > 0: lowerCamelCase_ = os.path.dirname(_lowerCAmelCase ) lowerCamelCase_ = str(hash(_lowerCAmelCase ) ) lowerCamelCase_ = filename[: max_length - len(_lowerCAmelCase ) - 8] + '...' + hashed_filename + '.lock' return os.path.join(_lowerCAmelCase , _lowerCAmelCase ) else: return path class _SCREAMING_SNAKE_CASE ( __snake_case ): def __init__( self , lowercase , lowercase=-1 , lowercase=None ) -> Optional[int]: from .file_utils import relative_to_absolute_path super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) lowerCamelCase_ = '\\\\?\\' + relative_to_absolute_path(self.lock_file ) def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: lowerCamelCase_ = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: lowerCamelCase_ = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: try: msvcrt.locking(_lowerCAmelCase , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(_lowerCAmelCase ) else: lowerCamelCase_ = fd return None def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self._lock_file_fd lowerCamelCase_ = None msvcrt.locking(_lowerCAmelCase , msvcrt.LK_UNLCK , 1 ) os.close(_lowerCAmelCase ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class _SCREAMING_SNAKE_CASE ( __snake_case ): def __init__( self , lowercase , lowercase=-1 , lowercase=None ) -> Dict: lowerCamelCase_ = os.statvfs(os.path.dirname(_lowerCAmelCase ) ).f_namemax super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = os.O_RDWR | os.O_CREAT | os.O_TRUNC lowerCamelCase_ = os.open(self._lock_file , _lowerCAmelCase ) try: fcntl.flock(_lowerCAmelCase , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(_lowerCAmelCase ) else: lowerCamelCase_ = fd return None def SCREAMING_SNAKE_CASE_( self ) -> Tuple: # Do not remove the lockfile: # # https://github.com/benediktschmitt/py-filelock/issues/31 # https://stackoverflow.com/questions/17708885/flock-removing-locked-file-without-race-condition lowerCamelCase_ = self._lock_file_fd lowerCamelCase_ = None fcntl.flock(_lowerCAmelCase , fcntl.LOCK_UN ) os.close(_lowerCAmelCase ) return None class _SCREAMING_SNAKE_CASE ( __snake_case ): def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: lowerCamelCase_ = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: lowerCamelCase_ = fd return None def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: os.close(self._lock_file_fd ) lowerCamelCase_ = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __A =None if msvcrt: __A =WindowsFileLock elif fcntl: __A =UnixFileLock else: __A =SoftFileLock if warnings is not None: warnings.warn('''only soft file lock is available''')
463
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
0
"""simple docstring""" import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class lowerCamelCase ( unittest.TestCase ): lowercase : Optional[Any] = JukeboxTokenizer lowercase : Tuple = { "artist": "Zac Brown Band", "genres": "Country", "lyrics": "I met a traveller from an antique land,\n Who said \"Two vast and trunkless legs of stone\n Stand in the desert. . . . Near them, on the sand,\n Half sunk a shattered visage lies, whose frown,\n And wrinkled lip, and sneer of cold command,\n Tell that its sculptor well those passions read\n Which yet survive, stamped on these lifeless things,\n The hand that mocked them, and the heart that fed;\n And on the pedestal, these words appear:\n My name is Ozymandias, King of Kings;\n Look on my Works, ye Mighty, and despair!\n Nothing beside remains. Round the decay\n Of that colossal Wreck, boundless and bare\n The lone and level sands stretch far away\n ", } @require_torch def a_ ( self ): import torch UpperCamelCase : List[Any] = JukeboxTokenizer.from_pretrained("""openai/jukebox-1b-lyrics""" ) UpperCamelCase : Optional[Any] = tokenizer(**self.metas )['input_ids'] # fmt: off UpperCamelCase : List[str] = [ torch.tensor([[ 0, 0, 0, 7169, 507, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def a_ ( self ): import torch UpperCamelCase : Optional[int] = JukeboxTokenizer.from_pretrained("""openai/jukebox-5b-lyrics""" ) UpperCamelCase : Tuple = tokenizer(**self.metas )['input_ids'] # fmt: off UpperCamelCase : Optional[Any] = [ torch.tensor([[ 0, 0, 0, 1069, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
499
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
0
'''simple docstring''' import os from datetime import datetime as dt from github import Github lowercase = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''enhancement''', '''new pipeline/model''', '''new scheduler''', '''wip''', ] def __A ( ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = Github(os.environ["GITHUB_TOKEN"] ) __SCREAMING_SNAKE_CASE : Tuple = g.get_repo("huggingface/diffusers" ) __SCREAMING_SNAKE_CASE : Optional[int] = repo.get_issues(state="open" ) for issue in open_issues: __SCREAMING_SNAKE_CASE : List[Any] = sorted(issue.get_comments() , key=lambda _SCREAMING_SNAKE_CASE : i.created_at , reverse=_SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE : Any = comments[0] if len(_SCREAMING_SNAKE_CASE ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="closed" ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="open" ) issue.remove_from_labels("stale" ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) issue.add_to_labels("stale" ) if __name__ == "__main__": main()
211
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _UpperCAmelCase : Tuple = logging.get_logger(__name__) def UpperCamelCase ( lowercase_ : str ) -> List[List[ImageInput]]: '''simple docstring''' if isinstance(lowercase_ , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowercase_ , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowercase_ ): return [[videos]] raise ValueError(f'Could not make batched video from {videos}' ) class __magic_name__ ( __snake_case ): UpperCamelCase__ = ["pixel_values"] def __init__( self , snake_case_ = True , snake_case_ = None , snake_case_ = PILImageResampling.BILINEAR , snake_case_ = True , snake_case_ = None , snake_case_ = True , snake_case_ = 1 / 2_55 , snake_case_ = True , snake_case_ = None , snake_case_ = None , **snake_case_ , ): super().__init__(**_lowerCAmelCase ) lowercase =size if size is not None else {'shortest_edge': 2_24} lowercase =get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) lowercase =crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} lowercase =get_size_dict(_lowerCAmelCase , param_name='''crop_size''' ) lowercase =do_resize lowercase =size lowercase =do_center_crop lowercase =crop_size lowercase =resample lowercase =do_rescale lowercase =rescale_factor lowercase =do_normalize lowercase =image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowercase =image_std if image_std is not None else IMAGENET_STANDARD_STD def _A( self , snake_case_ , snake_case_ , snake_case_ = PILImageResampling.BILINEAR , snake_case_ = None , **snake_case_ , ): lowercase =get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) if "shortest_edge" in size: lowercase =get_resize_output_image_size(_lowerCAmelCase , size['''shortest_edge'''] , default_to_square=_lowerCAmelCase ) elif "height" in size and "width" in size: lowercase =(size['height'], size['width']) else: raise ValueError(f'Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}' ) return resize(_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_ , snake_case_ = None , **snake_case_ , ): lowercase =get_size_dict(_lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'Size must have \'height\' and \'width\' as keys. Got {size.keys()}' ) return center_crop(_lowerCAmelCase , size=(size['''height'''], size['''width''']) , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_ , snake_case_ = None , **snake_case_ , ): return rescale(_lowerCAmelCase , scale=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ = None , **snake_case_ , ): return normalize(_lowerCAmelCase , mean=_lowerCAmelCase , std=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. lowercase =to_numpy_array(_lowerCAmelCase ) if do_resize: lowercase =self.resize(image=_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase ) if do_center_crop: lowercase =self.center_crop(_lowerCAmelCase , size=_lowerCAmelCase ) if do_rescale: lowercase =self.rescale(image=_lowerCAmelCase , scale=_lowerCAmelCase ) if do_normalize: lowercase =self.normalize(image=_lowerCAmelCase , mean=_lowerCAmelCase , std=_lowerCAmelCase ) lowercase =to_channel_dimension_format(_lowerCAmelCase , _lowerCAmelCase ) return image def _A( self , snake_case_ , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = ChannelDimension.FIRST , **snake_case_ , ): lowercase =do_resize if do_resize is not None else self.do_resize lowercase =resample if resample is not None else self.resample lowercase =do_center_crop if do_center_crop is not None else self.do_center_crop lowercase =do_rescale if do_rescale is not None else self.do_rescale lowercase =rescale_factor if rescale_factor is not None else self.rescale_factor lowercase =do_normalize if do_normalize is not None else self.do_normalize lowercase =image_mean if image_mean is not None else self.image_mean lowercase =image_std if image_std is not None else self.image_std lowercase =size if size is not None else self.size lowercase =get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) lowercase =crop_size if crop_size is not None else self.crop_size lowercase =get_size_dict(_lowerCAmelCase , param_name='''crop_size''' ) if not valid_images(_lowerCAmelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) lowercase =make_batched(_lowerCAmelCase ) lowercase =[ [ self._preprocess_image( image=_lowerCAmelCase , do_resize=_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase , do_center_crop=_lowerCAmelCase , crop_size=_lowerCAmelCase , do_rescale=_lowerCAmelCase , rescale_factor=_lowerCAmelCase , do_normalize=_lowerCAmelCase , image_mean=_lowerCAmelCase , image_std=_lowerCAmelCase , data_format=_lowerCAmelCase , ) for img in video ] for video in videos ] lowercase ={'pixel_values': videos} return BatchFeature(data=_lowerCAmelCase , tensor_type=_lowerCAmelCase )
72
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
0
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging UpperCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = R'\w+[.]\d+' UpperCAmelCase__ = re.findall(__A, __A ) for pat in pats: UpperCAmelCase__ = key.replace(__A, "_".join(pat.split("." ) ) ) return key def lowerCAmelCase_ ( __A, __A, __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = pt_tuple_key[:-1] + ('scale',) if ( any("norm" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): UpperCAmelCase__ = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: UpperCAmelCase__ = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: UpperCAmelCase__ = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer UpperCAmelCase__ = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: UpperCAmelCase__ = pt_tensor.transpose(2, 3, 1, 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer UpperCAmelCase__ = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": UpperCAmelCase__ = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight UpperCAmelCase__ = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias UpperCAmelCase__ = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCAmelCase_ ( __A, __A, __A=42 ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params UpperCAmelCase__ = flax_model.init_weights(PRNGKey(__A ) ) UpperCAmelCase__ = flatten_dict(__A ) UpperCAmelCase__ = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): UpperCAmelCase__ = rename_key(__A ) UpperCAmelCase__ = tuple(renamed_pt_key.split("." ) ) # Correctly rename weight parameters UpperCAmelCase__ = rename_key_and_reshape_tensor(__A, __A, __A ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown UpperCAmelCase__ = jnp.asarray(__A ) return unflatten_dict(__A )
486
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
0
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase : List[Any] ={ "configuration_efficientnet": [ "EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientNetConfig", "EfficientNetOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict =["EfficientNetImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : str =[ "EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientNetForImageClassification", "EfficientNetModel", "EfficientNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys _lowercase : Union[str, Any] =_LazyModule(__name__, globals()["__file__"], _import_structure)
136
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : str = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Dict = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys lowercase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
327
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
0
'''simple docstring''' from __future__ import annotations import requests lowerCamelCase :Tuple = set( '''approved_at_utc approved_by author_flair_background_color\nauthor_flair_css_class author_flair_richtext author_flair_template_id author_fullname\nauthor_premium can_mod_post category clicked content_categories created_utc downs\nedited gilded gildings hidden hide_score is_created_from_ads_ui is_meta\nis_original_content is_reddit_media_domain is_video link_flair_css_class\nlink_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title\nname permalink pwls quarantine saved score secure_media secure_media_embed selftext\nsubreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type\ntotal_awards_received ups upvote_ratio url user_reports'''.split() ) def a ( lowerCamelCase__ , lowerCamelCase__ = 1 , lowerCamelCase__ = "new" , lowerCamelCase__ = None ): '''simple docstring''' A_ : Dict = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(lowerCamelCase__ ) - valid_terms ) ): A_ : List[str] = f'Invalid search term: {invalid_search_terms}' raise ValueError(lowerCamelCase__ ) A_ : Dict = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 4_29: raise requests.HTTPError A_ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(lowerCamelCase__ )} A_ : List[Any] = {} for id_ in range(lowerCamelCase__ ): A_ : List[str] = { item: data['data']['children'][id_]['data'][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
667
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() __A : Union[str, Any] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> List[Any]: '''simple docstring''' lowerCAmelCase : Optional[Any] = SwinConfig( embed_dim=192, depths=(2, 2, 18, 2), num_heads=(6, 12, 24, 48), window_size=12, out_features=['stage2', 'stage3', 'stage4'], ) lowerCAmelCase : Any = DetaConfig( backbone_config=_UpperCAmelCase, num_queries=900, encoder_ffn_dim=2_048, decoder_ffn_dim=2_048, num_feature_levels=5, assign_first_stage=_UpperCAmelCase, with_box_refine=_UpperCAmelCase, two_stage=_UpperCAmelCase, ) # set labels lowerCAmelCase : int = 'huggingface/label-files' if "o365" in model_name: lowerCAmelCase : Any = 366 lowerCAmelCase : Dict = 'object365-id2label.json' else: lowerCAmelCase : str = 91 lowerCAmelCase : Union[str, Any] = 'coco-detection-id2label.json' lowerCAmelCase : Optional[int] = num_labels lowerCAmelCase : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase, _UpperCAmelCase, repo_type='dataset' ) ), 'r' ) ) lowerCAmelCase : List[str] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCAmelCase : Any = idalabel lowerCAmelCase : Union[str, Any] = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase : Union[str, Any] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.patch_embed.proj.weight', 'model.backbone.model.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.0.body.patch_embed.proj.bias', 'model.backbone.model.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.0.body.patch_embed.norm.weight', 'model.backbone.model.embeddings.norm.weight') ) rename_keys.append(('backbone.0.body.patch_embed.norm.bias', 'model.backbone.model.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.norm1.weight", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.norm1.bias", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.norm2.weight", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.norm2.bias", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias", f"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias") ) if i < 3: rename_keys.append((f"backbone.0.body.layers.{i}.downsample.reduction.weight", f"model.backbone.model.encoder.layers.{i}.downsample.reduction.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.downsample.norm.weight", f"model.backbone.model.encoder.layers.{i}.downsample.norm.weight") ) rename_keys.append((f"backbone.0.body.layers.{i}.downsample.norm.bias", f"model.backbone.model.encoder.layers.{i}.downsample.norm.bias") ) rename_keys.append(('backbone.0.body.norm1.weight', 'model.backbone.model.hidden_states_norms.stage2.weight') ) rename_keys.append(('backbone.0.body.norm1.bias', 'model.backbone.model.hidden_states_norms.stage2.bias') ) rename_keys.append(('backbone.0.body.norm2.weight', 'model.backbone.model.hidden_states_norms.stage3.weight') ) rename_keys.append(('backbone.0.body.norm2.bias', 'model.backbone.model.hidden_states_norms.stage3.bias') ) rename_keys.append(('backbone.0.body.norm3.weight', 'model.backbone.model.hidden_states_norms.stage4.weight') ) rename_keys.append(('backbone.0.body.norm3.bias', 'model.backbone.model.hidden_states_norms.stage4.bias') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight", f"model.encoder.layers.{i}.self_attn.sampling_offsets.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias", f"model.encoder.layers.{i}.self_attn.sampling_offsets.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.attention_weights.weight", f"model.encoder.layers.{i}.self_attn.attention_weights.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.attention_weights.bias", f"model.encoder.layers.{i}.self_attn.attention_weights.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.value_proj.weight", f"model.encoder.layers.{i}.self_attn.value_proj.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.value_proj.bias", f"model.encoder.layers.{i}.self_attn.value_proj.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.output_proj.weight", f"model.encoder.layers.{i}.self_attn.output_proj.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.self_attn.output_proj.bias", f"model.encoder.layers.{i}.self_attn.output_proj.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm1.weight", f"model.encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm1.bias", f"model.encoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"model.encoder.layers.{i}.fc1.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"model.encoder.layers.{i}.fc1.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"model.encoder.layers.{i}.fc2.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"model.encoder.layers.{i}.fc2.bias") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.weight", f"model.encoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"model.encoder.layers.{i}.final_layer_norm.bias") ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight", f"model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias", f"model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.attention_weights.weight", f"model.decoder.layers.{i}.encoder_attn.attention_weights.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.attention_weights.bias", f"model.decoder.layers.{i}.encoder_attn.attention_weights.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.value_proj.weight", f"model.decoder.layers.{i}.encoder_attn.value_proj.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.value_proj.bias", f"model.decoder.layers.{i}.encoder_attn.value_proj.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.output_proj.weight", f"model.decoder.layers.{i}.encoder_attn.output_proj.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.cross_attn.output_proj.bias", f"model.decoder.layers.{i}.encoder_attn.output_proj.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm1.weight", f"model.decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm1.bias", f"model.decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"model.decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"model.decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm2.weight", f"model.decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm2.bias", f"model.decoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"model.decoder.layers.{i}.fc1.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"model.decoder.layers.{i}.fc1.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"model.decoder.layers.{i}.fc2.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"model.decoder.layers.{i}.fc2.bias") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.weight", f"model.decoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"model.decoder.layers.{i}.final_layer_norm.bias") ) # fmt: on return rename_keys def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> List[str]: '''simple docstring''' lowerCAmelCase : Any = dct.pop(_UpperCAmelCase ) lowerCAmelCase : List[str] = val def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase : Optional[Any] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): lowerCAmelCase : Union[str, Any] = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) lowerCAmelCase : str = state_dict.pop(f"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight" ) lowerCAmelCase : Optional[Any] = state_dict.pop(f"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase : List[str] = in_proj_weight[:dim, :] lowerCAmelCase : List[str] = in_proj_bias[: dim] lowerCAmelCase : str = in_proj_weight[ dim : dim * 2, : ] lowerCAmelCase : Union[str, Any] = in_proj_bias[ dim : dim * 2 ] lowerCAmelCase : Dict = in_proj_weight[ -dim :, : ] lowerCAmelCase : Union[str, Any] = in_proj_bias[-dim :] # fmt: on def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase : str = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention lowerCAmelCase : Dict = state_dict.pop(f"transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) lowerCAmelCase : str = state_dict.pop(f"transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase : List[Any] = in_proj_weight[:hidden_size, :] lowerCAmelCase : Optional[Any] = in_proj_bias[:hidden_size] lowerCAmelCase : List[str] = in_proj_weight[ hidden_size : hidden_size * 2, : ] lowerCAmelCase : str = in_proj_bias[hidden_size : hidden_size * 2] lowerCAmelCase : str = in_proj_weight[-hidden_size:, :] lowerCAmelCase : Dict = in_proj_bias[-hidden_size:] def SCREAMING_SNAKE_CASE__ ( ) -> Optional[int]: '''simple docstring''' lowerCAmelCase : str = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCAmelCase : List[Any] = Image.open(requests.get(_UpperCAmelCase, stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase : Dict = get_deta_config(_UpperCAmelCase ) # load original state dict if model_name == "deta-swin-large": lowerCAmelCase : int = hf_hub_download(repo_id='nielsr/deta-checkpoints', filename='adet_swin_ft.pth' ) elif model_name == "deta-swin-large-o365": lowerCAmelCase : Dict = hf_hub_download(repo_id='jozhang97/deta-swin-l-o365', filename='deta_swin_pt_o365.pth' ) else: raise ValueError(f"Model name {model_name} not supported" ) lowerCAmelCase : str = torch.load(_UpperCAmelCase, map_location='cpu' )['model'] # original state dict for name, param in state_dict.items(): print(_UpperCAmelCase, param.shape ) # rename keys lowerCAmelCase : Tuple = create_rename_keys(_UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) read_in_swin_q_k_v(_UpperCAmelCase, config.backbone_config ) read_in_decoder_q_k_v(_UpperCAmelCase, _UpperCAmelCase ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: lowerCAmelCase : int = state_dict.pop(_UpperCAmelCase ) lowerCAmelCase : List[str] = val if "input_proj" in key: lowerCAmelCase : Any = state_dict.pop(_UpperCAmelCase ) lowerCAmelCase : int = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: lowerCAmelCase : Union[str, Any] = state_dict.pop(_UpperCAmelCase ) lowerCAmelCase : Dict = val # finally, create HuggingFace model and load state dict lowerCAmelCase : Optional[Any] = DetaForObjectDetection(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() lowerCAmelCase : Tuple = 'cuda' if torch.cuda.is_available() else 'cpu' model.to(_UpperCAmelCase ) # load image processor lowerCAmelCase : Optional[int] = DetaImageProcessor(format='coco_detection' ) # verify our conversion on image lowerCAmelCase : int = prepare_img() lowerCAmelCase : Tuple = processor(images=_UpperCAmelCase, return_tensors='pt' ) lowerCAmelCase : Union[str, Any] = encoding['pixel_values'] lowerCAmelCase : Any = model(pixel_values.to(_UpperCAmelCase ) ) # verify logits print('Logits:', outputs.logits[0, :3, :3] ) print('Boxes:', outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": lowerCAmelCase : Tuple = torch.tensor( [[-7.6_3_0_8, -2.8_4_8_5, -5.3_7_3_7], [-7.2_0_3_7, -4.5_5_0_5, -4.8_0_2_7], [-7.2_9_4_3, -4.2_6_1_1, -4.6_6_1_7]] ) lowerCAmelCase : Dict = torch.tensor([[0.4_9_8_7, 0.4_9_6_9, 0.9_9_9_9], [0.2_5_4_9, 0.5_4_9_8, 0.4_8_0_5], [0.5_4_9_8, 0.2_7_5_7, 0.0_5_6_9]] ) elif model_name == "deta-swin-large-o365": lowerCAmelCase : Union[str, Any] = torch.tensor( [[-8.0_1_2_2, -3.5_7_2_0, -4.9_7_1_7], [-8.1_5_4_7, -3.6_8_8_6, -4.6_3_8_9], [-7.6_6_1_0, -3.6_1_9_4, -5.0_1_3_4]] ) lowerCAmelCase : Dict = torch.tensor([[0.2_5_2_3, 0.5_5_4_9, 0.4_8_8_1], [0.7_7_1_5, 0.4_1_4_9, 0.4_6_0_1], [0.5_5_0_3, 0.2_7_5_3, 0.0_5_7_5]] ) assert torch.allclose(outputs.logits[0, :3, :3], expected_logits.to(_UpperCAmelCase ), atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3], expected_boxes.to(_UpperCAmelCase ), atol=1e-4 ) print('Everything ok!' ) if pytorch_dump_folder_path: # Save model and processor logger.info(f"Saving PyTorch model and processor to {pytorch_dump_folder_path}..." ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) # Push to hub if push_to_hub: print('Pushing model and processor to hub...' ) model.push_to_hub(f"jozhang97/{model_name}" ) processor.push_to_hub(f"jozhang97/{model_name}" ) if __name__ == "__main__": __A : List[str] = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __A : List[Any] = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
343
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase = logging.get_logger(__name__) def a ( a ) ->Tuple: '''simple docstring''' if "resnet-50" in model_name: SCREAMING_SNAKE_CASE = ResNetConfig.from_pretrained('''microsoft/resnet-50''' ) elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE = ResNetConfig.from_pretrained('''microsoft/resnet-101''' ) else: raise ValueError('''Model name should include either resnet50 or resnet101''' ) SCREAMING_SNAKE_CASE = DetrConfig(use_timm_backbone=a , backbone_config=a ) # set label attributes SCREAMING_SNAKE_CASE = 'panoptic' in model_name if is_panoptic: SCREAMING_SNAKE_CASE = 250 else: SCREAMING_SNAKE_CASE = 91 SCREAMING_SNAKE_CASE = 'huggingface/label-files' SCREAMING_SNAKE_CASE = 'coco-detection-id2label.json' SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) SCREAMING_SNAKE_CASE = {int(a ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = idalabel SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config, is_panoptic def a ( a ) ->Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = [] # stem # fmt: off rename_keys.append(('''backbone.0.body.conv1.weight''', '''backbone.conv_encoder.model.embedder.embedder.convolution.weight''') ) rename_keys.append(('''backbone.0.body.bn1.weight''', '''backbone.conv_encoder.model.embedder.embedder.normalization.weight''') ) rename_keys.append(('''backbone.0.body.bn1.bias''', '''backbone.conv_encoder.model.embedder.embedder.normalization.bias''') ) rename_keys.append(('''backbone.0.body.bn1.running_mean''', '''backbone.conv_encoder.model.embedder.embedder.normalization.running_mean''') ) rename_keys.append(('''backbone.0.body.bn1.running_var''', '''backbone.conv_encoder.model.embedder.embedder.normalization.running_var''') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var""", ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var""", ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""") ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""") ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ] ) return rename_keys def a ( a , a , a ) ->Any: '''simple docstring''' SCREAMING_SNAKE_CASE = state_dict.pop(a ) SCREAMING_SNAKE_CASE = val def a ( a , a=False ) ->Any: '''simple docstring''' SCREAMING_SNAKE_CASE = '' if is_panoptic: SCREAMING_SNAKE_CASE = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) SCREAMING_SNAKE_CASE = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE = in_proj_bias[:256] SCREAMING_SNAKE_CASE = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE = in_proj_bias[256:512] SCREAMING_SNAKE_CASE = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) SCREAMING_SNAKE_CASE = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE = in_proj_bias[:256] SCREAMING_SNAKE_CASE = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE = in_proj_bias[256:512] SCREAMING_SNAKE_CASE = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE = state_dict.pop( F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight""" ) SCREAMING_SNAKE_CASE = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE = in_proj_bias_cross_attn[-256:] def a ( ) ->str: '''simple docstring''' SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def a ( a , a=None , a=False ) ->Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = get_detr_config(a ) # load original model from torch hub SCREAMING_SNAKE_CASE = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F"""Converting model {model_name}...""" ) SCREAMING_SNAKE_CASE = torch.hub.load('''facebookresearch/detr''' , model_name_to_original_name[model_name] , pretrained=a ).eval() SCREAMING_SNAKE_CASE = detr.state_dict() # rename keys for src, dest in create_rename_keys(a ): if is_panoptic: SCREAMING_SNAKE_CASE = 'detr.' + src rename_key(a , a , a ) # query, key and value matrices need special treatment read_in_q_k_v(a , is_panoptic=a ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('''detr''' ) and not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ) ): SCREAMING_SNAKE_CASE = state_dict.pop(a ) SCREAMING_SNAKE_CASE = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE = state_dict.pop(a ) SCREAMING_SNAKE_CASE = val elif key.startswith('''bbox_attention''' ) or key.startswith('''mask_head''' ): continue else: SCREAMING_SNAKE_CASE = state_dict.pop(a ) SCREAMING_SNAKE_CASE = val else: if not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ): SCREAMING_SNAKE_CASE = state_dict.pop(a ) SCREAMING_SNAKE_CASE = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE = DetrForSegmentation(a ) if is_panoptic else DetrForObjectDetection(a ) model.load_state_dict(a ) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE = 'coco_panoptic' if is_panoptic else 'coco_detection' SCREAMING_SNAKE_CASE = DetrImageProcessor(format=a ) SCREAMING_SNAKE_CASE = processor(images=prepare_img() , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE = encoding['pixel_values'] SCREAMING_SNAKE_CASE = detr(a ) SCREAMING_SNAKE_CASE = model(a ) assert torch.allclose(outputs.logits , original_outputs['''pred_logits'''] , atol=1E-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['''pred_boxes'''] , atol=1E-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['''pred_masks'''] , atol=1E-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) processor.save_pretrained(a ) if push_to_hub: # Upload model and image processor to the hub logger.info('''Uploading PyTorch model and image processor to the hub...''' ) model.push_to_hub(F"""nielsr/{model_name}""" ) processor.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') __lowerCAmelCase = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
201
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
0
'''simple docstring''' import os from datetime import datetime as dt from github import Github __SCREAMING_SNAKE_CASE :List[str] = [ '''good first issue''', '''feature request''', '''wip''', ] def UpperCAmelCase_ ( ) -> List[str]: '''simple docstring''' _UpperCAmelCase = Github(os.environ["GITHUB_TOKEN"] ) _UpperCAmelCase = g.get_repo("huggingface/accelerate" ) _UpperCAmelCase = repo.get_issues(state="open" ) for issue in open_issues: _UpperCAmelCase = sorted([comment for comment in issue.get_comments()] , key=lambda __lowercase : i.created_at , reverse=__lowercase ) _UpperCAmelCase = comments[0] if len(__lowercase ) > 0 else None _UpperCAmelCase = dt.utcnow() _UpperCAmelCase = (current_time - issue.updated_at).days _UpperCAmelCase = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state="closed" ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
236
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __A =logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE ( __snake_case ): def __init__( self , *lowercase , **lowercase ) -> str: warnings.warn( "The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use BeitImageProcessor instead." , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
463
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __A : List[str] = { '''configuration_efficientformer''': [ '''EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EfficientFormerConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = ['''EfficientFormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ '''EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EfficientFormerForImageClassification''', '''EfficientFormerForImageClassificationWithTeacher''', '''EfficientFormerModel''', '''EfficientFormerPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ '''TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFEfficientFormerForImageClassification''', '''TFEfficientFormerForImageClassificationWithTeacher''', '''TFEfficientFormerModel''', '''TFEfficientFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys __A : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
499
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
0
'''simple docstring''' from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name class __lowerCamelCase ( __snake_case ): '''simple docstring''' def __init__( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __SCREAMING_SNAKE_CASE : Dict = ( f'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' f' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate("steps_offset!=1" , "1.0.0" , _lowerCAmelCase , standard_warn=_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : List[str] = dict(scheduler.config ) __SCREAMING_SNAKE_CASE : Any = 1 __SCREAMING_SNAKE_CASE : List[Any] = FrozenDict(_lowerCAmelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __SCREAMING_SNAKE_CASE : Any = ( f'The configuration file of this scheduler: {scheduler} has not set the configuration' ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate("skip_prk_steps not set" , "1.0.0" , _lowerCAmelCase , standard_warn=_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : str = dict(scheduler.config ) __SCREAMING_SNAKE_CASE : int = True __SCREAMING_SNAKE_CASE : Union[str, Any] = FrozenDict(_lowerCAmelCase ) if safety_checker is None: logger.warning( f'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=_lowerCAmelCase , segmentation_processor=_lowerCAmelCase , vae=_lowerCAmelCase , text_encoder=_lowerCAmelCase , tokenizer=_lowerCAmelCase , unet=_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , ) def a_ ( self , a__ = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __SCREAMING_SNAKE_CASE : str = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_lowerCAmelCase ) def a_ ( self ): self.enable_attention_slicing(_lowerCAmelCase ) def a_ ( self ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __SCREAMING_SNAKE_CASE : Tuple = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(_lowerCAmelCase , _lowerCAmelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a_ ( self ): if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowerCAmelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , a__ , a__ , a__ , a__ = 512 , a__ = 512 , a__ = 50 , a__ = 7.5 , a__ = None , a__ = 1 , a__ = 0.0 , a__ = None , a__ = None , a__ = "pil" , a__ = True , a__ = None , a__ = 1 , **a__ , ): __SCREAMING_SNAKE_CASE : str = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __SCREAMING_SNAKE_CASE : Any = self.segmentation_model(**_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : str = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __SCREAMING_SNAKE_CASE : Optional[int] = self.numpy_to_pil(_lowerCAmelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __SCREAMING_SNAKE_CASE : str = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=_lowerCAmelCase , image=_lowerCAmelCase , mask_image=_lowerCAmelCase , height=_lowerCAmelCase , width=_lowerCAmelCase , num_inference_steps=_lowerCAmelCase , guidance_scale=_lowerCAmelCase , negative_prompt=_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase , eta=_lowerCAmelCase , generator=_lowerCAmelCase , latents=_lowerCAmelCase , output_type=_lowerCAmelCase , return_dict=_lowerCAmelCase , callback=_lowerCAmelCase , callback_steps=_lowerCAmelCase , )
211
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
0
'''simple docstring''' import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) # pylint: disable=invalid-name class __magic_name__ ( __snake_case ): def __init__( self , snake_case_ , snake_case_=7_68 ): super().__init__(_lowerCAmelCase ) lowercase =proj_size lowercase =CLIPVisionModel(_lowerCAmelCase ) lowercase =PaintByExampleMapper(_lowerCAmelCase ) lowercase =nn.LayerNorm(config.hidden_size ) lowercase =nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling lowercase =nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def _A( self , snake_case_ , snake_case_=False ): lowercase =self.model(pixel_values=_lowerCAmelCase ) lowercase =clip_output.pooler_output lowercase =self.mapper(latent_states[:, None] ) lowercase =self.final_layer_norm(_lowerCAmelCase ) lowercase =self.proj_out(_lowerCAmelCase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class __magic_name__ ( nn.Module ): def __init__( self , snake_case_ ): super().__init__() lowercase =(config.num_hidden_layers + 1) // 5 lowercase =config.hidden_size lowercase =1 lowercase =nn.ModuleList( [ BasicTransformerBlock(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , activation_fn='''gelu''' , attention_bias=_lowerCAmelCase ) for _ in range(_lowerCAmelCase ) ] ) def _A( self , snake_case_ ): for block in self.blocks: lowercase =block(_lowerCAmelCase ) return hidden_states
72
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
0
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets UpperCamelCase__ = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' UpperCamelCase__ = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' UpperCamelCase__ = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: \"c\" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A ( datasets.Metric ): def lowercase_ (self : int ) -> List[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/krishnap25/mauve" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/krishnap25/mauve"] , reference_urls=[ "https://arxiv.org/abs/2102.01454", "https://github.com/krishnap25/mauve", ] , ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any=None , __UpperCAmelCase : Any=None , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : Any=None , __UpperCAmelCase : Dict="auto" , __UpperCAmelCase : Optional[int]=-1 , __UpperCAmelCase : Optional[Any]=0.9 , __UpperCAmelCase : Optional[int]=5 , __UpperCAmelCase : List[str]=5_0_0 , __UpperCAmelCase : Optional[int]="gpt2-large" , __UpperCAmelCase : Optional[int]=-1 , __UpperCAmelCase : Any=1_0_2_4 , __UpperCAmelCase : Union[str, Any]=2_5 , __UpperCAmelCase : Optional[Any]=5 , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Tuple=2_5 , ) -> Tuple: """simple docstring""" UpperCAmelCase__ = compute_mauve( p_text=_lowerCAmelCase , q_text=_lowerCAmelCase , p_features=_lowerCAmelCase , q_features=_lowerCAmelCase , p_tokens=_lowerCAmelCase , q_tokens=_lowerCAmelCase , num_buckets=_lowerCAmelCase , pca_max_data=_lowerCAmelCase , kmeans_explained_var=_lowerCAmelCase , kmeans_num_redo=_lowerCAmelCase , kmeans_max_iter=_lowerCAmelCase , featurize_model_name=_lowerCAmelCase , device_id=_lowerCAmelCase , max_text_length=_lowerCAmelCase , divergence_curve_discretization_size=_lowerCAmelCase , mauve_scaling_factor=_lowerCAmelCase , verbose=_lowerCAmelCase , seed=_lowerCAmelCase , ) return out
486
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
0
from __future__ import annotations from math import pi, sqrt def lowerCAmelCase_ ( _lowercase : Union[str, Any] , _lowercase : Optional[Any]) -> tuple: """simple docstring""" if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""") elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""") else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance)))), ) if __name__ == "__main__": import doctest doctest.testmod()
136
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
import functools from typing import Any def UpperCAmelCase_ (_lowerCAmelCase : int , _lowerCAmelCase : List[str] ): # Validation if not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or len(_lowerCAmelCase ) == 0: raise ValueError("the string should be not empty string" ) if not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or not all( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and len(_lowerCAmelCase ) > 0 for item in words ): raise ValueError("the words should be a list of non-empty strings" ) # Build trie __UpperCamelCase : dict[str, Any] = {} __UpperCamelCase : Any = 'WORD_KEEPER' for word in words: __UpperCamelCase : List[str] = trie for c in word: if c not in trie_node: __UpperCamelCase : List[Any] = {} __UpperCamelCase : int = trie_node[c] __UpperCamelCase : Tuple = True __UpperCamelCase : Optional[Any] = len(_lowerCAmelCase ) # Dynamic programming method @functools.cache def is_breakable(_lowerCAmelCase : Optional[Any] ) -> bool: if index == len_string: return True __UpperCamelCase : Union[str, Any] = trie for i in range(_lowerCAmelCase , _lowerCAmelCase ): __UpperCamelCase : Tuple = trie_node.get(string[i] , _lowerCAmelCase ) if trie_node is None: return False if trie_node.get(_lowerCAmelCase , _lowerCAmelCase ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
327
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
0
'''simple docstring''' import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu lowerCamelCase :Dict = get_tests_dir() + '''/test_data/fsmt/fsmt_val_data.json''' with io.open(filename, '''r''', encoding='''utf-8''') as f: lowerCamelCase :Optional[int] = json.load(f) @require_torch class _lowerCAmelCase ( unittest.TestCase ): def _a (self , lowercase ): return FSMTTokenizer.from_pretrained(_lowerCAmelCase ) def _a (self , lowercase ): A_ : Tuple = FSMTForConditionalGeneration.from_pretrained(_lowerCAmelCase ).to(_lowerCAmelCase ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ["""en-ru""", 26.0], ["""ru-en""", 22.0], ["""en-de""", 22.0], ["""de-en""", 29.0], ] ) @slow def _a (self , lowercase , lowercase ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality A_ : Optional[Any] = F'facebook/wmt19-{pair}' A_ : Any = self.get_tokenizer(_lowerCAmelCase ) A_ : Union[str, Any] = self.get_model(_lowerCAmelCase ) A_ : Any = bleu_data[pair]['src'] A_ : Tuple = bleu_data[pair]['tgt'] A_ : Tuple = tokenizer(_lowerCAmelCase , return_tensors="""pt""" , truncation=_lowerCAmelCase , padding="""longest""" ).to(_lowerCAmelCase ) A_ : List[Any] = model.generate( input_ids=batch.input_ids , num_beams=8 , ) A_ : Union[str, Any] = tokenizer.batch_decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) A_ : Optional[Any] = calculate_bleu(_lowerCAmelCase , _lowerCAmelCase ) print(_lowerCAmelCase ) self.assertGreaterEqual(scores["""bleu"""] , _lowerCAmelCase )
667
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
0
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __A ( __snake_case ): def __init__( self : Optional[int] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any]=13 , UpperCAmelCase_ : str=7 , UpperCAmelCase_ : Any=True , UpperCAmelCase_ : str=True , UpperCAmelCase_ : int=True , UpperCAmelCase_ : Any=True , UpperCAmelCase_ : Optional[Any]=99 , UpperCAmelCase_ : Tuple=32 , UpperCAmelCase_ : Dict=5 , UpperCAmelCase_ : Tuple=4 , UpperCAmelCase_ : str=37 , UpperCAmelCase_ : str="gelu" , UpperCAmelCase_ : int=0.1 , UpperCAmelCase_ : List[Any]=0.1 , UpperCAmelCase_ : Tuple=512 , UpperCAmelCase_ : List[str]=16 , UpperCAmelCase_ : str=2 , UpperCAmelCase_ : Tuple=0.02 , UpperCAmelCase_ : str=False , UpperCAmelCase_ : List[str]=True , UpperCAmelCase_ : Optional[Any]="None" , UpperCAmelCase_ : Optional[Any]=3 , UpperCAmelCase_ : Union[str, Any]=4 , UpperCAmelCase_ : Optional[Any]=None , ): lowerCAmelCase : Any = parent lowerCAmelCase : Optional[Any] = batch_size lowerCAmelCase : Union[str, Any] = seq_length lowerCAmelCase : List[Any] = is_training lowerCAmelCase : Any = use_input_mask lowerCAmelCase : Dict = use_token_type_ids lowerCAmelCase : Tuple = use_labels lowerCAmelCase : Union[str, Any] = vocab_size lowerCAmelCase : Optional[Any] = hidden_size lowerCAmelCase : Optional[int] = num_hidden_layers lowerCAmelCase : List[str] = num_attention_heads lowerCAmelCase : Tuple = intermediate_size lowerCAmelCase : int = hidden_act lowerCAmelCase : str = hidden_dropout_prob lowerCAmelCase : Any = attention_probs_dropout_prob lowerCAmelCase : Dict = max_position_embeddings lowerCAmelCase : List[Any] = type_vocab_size lowerCAmelCase : int = type_sequence_label_size lowerCAmelCase : Dict = initializer_range lowerCAmelCase : Dict = num_labels lowerCAmelCase : Tuple = num_choices lowerCAmelCase : Optional[int] = relative_attention lowerCAmelCase : int = position_biased_input lowerCAmelCase : int = pos_att_type lowerCAmelCase : str = scope def lowercase__ ( self : str ): lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : Optional[int] = None if self.use_input_mask: lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowerCAmelCase : str = None if self.use_token_type_ids: lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase : Optional[Any] = None lowerCAmelCase : List[str] = None lowerCAmelCase : Any = None if self.use_labels: lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase : Optional[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self : int ): return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def lowercase__ ( self : List[str] , UpperCAmelCase_ : Optional[Any] ): self.parent.assertListEqual(list(result.loss.size() ) , [] ) def lowercase__ ( self : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Any ): lowerCAmelCase : Optional[Any] = DebertaVaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : Tuple = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase )[0] lowerCAmelCase : int = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase )[0] lowerCAmelCase : Tuple = model(_lowerCAmelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def lowercase__ ( self : Optional[int] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Any ): lowerCAmelCase : Optional[Any] = DebertaVaForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : str = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : Any , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Tuple ): lowerCAmelCase : Dict = self.num_labels lowerCAmelCase : Optional[int] = DebertaVaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : Optional[int] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(_lowerCAmelCase ) def lowercase__ ( self : Any , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any] ): lowerCAmelCase : Any = self.num_labels lowerCAmelCase : List[str] = DebertaVaForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : int = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase__ ( self : Any , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] ): lowerCAmelCase : Union[str, Any] = DebertaVaForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : Optional[int] = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase__ ( self : Dict , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : str ): lowerCAmelCase : Tuple = DebertaVaForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCAmelCase : Any = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : List[str] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : int = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : int = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : Dict = self.prepare_config_and_inputs() ( lowerCAmelCase ) : int = config_and_inputs lowerCAmelCase : Any = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __A ( __snake_case , __snake_case , unittest.TestCase ): lowerCAmelCase_ : Union[str, Any] = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) lowerCAmelCase_ : Optional[int] = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ : List[Any] = True lowerCAmelCase_ : List[str] = False lowerCAmelCase_ : Any = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : Optional[Any] = False def lowercase__ ( self : Optional[int] ): lowerCAmelCase : Any = DebertaVaModelTester(self ) lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def lowercase__ ( self : List[str] ): self.config_tester.run_common_tests() def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*_lowerCAmelCase ) def lowercase__ ( self : List[Any] ): lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*_lowerCAmelCase ) def lowercase__ ( self : Optional[int] ): lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*_lowerCAmelCase ) def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*_lowerCAmelCase ) def lowercase__ ( self : Any ): lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*_lowerCAmelCase ) def lowercase__ ( self : int ): lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*_lowerCAmelCase ) @slow def lowercase__ ( self : int ): for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase : Any = DebertaVaModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): @unittest.skip(reason='Model not available yet' ) def lowercase__ ( self : str ): pass @slow def lowercase__ ( self : List[Any] ): lowerCAmelCase : int = DebertaVaModel.from_pretrained('microsoft/deberta-v2-xlarge' ) lowerCAmelCase : Optional[Any] = torch.tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) lowerCAmelCase : List[str] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase : Any = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] # compare the actual values for a slice. lowerCAmelCase : Optional[int] = torch.tensor( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1E-4 ) , f"{output[:, 1:4, 1:4]}" )
343
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'SwiftFormerOnnxConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'SwiftFormerForImageClassification', 'SwiftFormerModel', 'SwiftFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
201
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class A_ : def __init__( self : List[str] , snake_case_ : Optional[Any] ): _UpperCAmelCase = str(id_ ) _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = [] _UpperCAmelCase = {} # {vertex:distance} def __lt__( self : Any , snake_case_ : Union[str, Any] ): return self.key < other.key def __repr__( self : List[Any] ): return self.id def lowercase ( self : List[str] , snake_case_ : str ): self.neighbors.append(_lowerCAmelCase ) def lowercase ( self : Tuple , snake_case_ : Optional[Any] , snake_case_ : List[Any] ): _UpperCAmelCase = weight def UpperCAmelCase_ ( __lowercase : str , __lowercase : Tuple , __lowercase : int , __lowercase : Dict ) -> List[Any]: '''simple docstring''' graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __lowercase ) graph[b - 1].add_edge(graph[a - 1] , __lowercase ) def UpperCAmelCase_ ( __lowercase : Dict , __lowercase : Any ) -> list: '''simple docstring''' _UpperCAmelCase = [] for u in graph: _UpperCAmelCase = math.inf _UpperCAmelCase = None _UpperCAmelCase = 0 _UpperCAmelCase = graph[:] while q: _UpperCAmelCase = min(__lowercase ) q.remove(__lowercase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _UpperCAmelCase = u _UpperCAmelCase = u.edges[v.id] for i in range(1 , len(__lowercase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCAmelCase_ ( __lowercase : Tuple , __lowercase : Union[str, Any] ) -> Iterator[tuple]: '''simple docstring''' for u in graph: _UpperCAmelCase = math.inf _UpperCAmelCase = None _UpperCAmelCase = 0 _UpperCAmelCase = list(__lowercase ) hq.heapify(__lowercase ) while h: _UpperCAmelCase = hq.heappop(__lowercase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _UpperCAmelCase = u _UpperCAmelCase = u.edges[v.id] hq.heapify(__lowercase ) for i in range(1 , len(__lowercase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCAmelCase_ ( ) -> None: '''simple docstring''' pass if __name__ == "__main__": import doctest doctest.testmod()
236
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
0
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): # Load checkpoint lowerCamelCase_ = torch.load(lowerCamelCase__ , map_location="cpu" ) lowerCamelCase_ = chkpt['model'] # We have the base model one level deeper than the original XLM repository lowerCamelCase_ = {} for k, v in state_dict.items(): if "pred_layer" in k: lowerCamelCase_ = v else: lowerCamelCase_ = v lowerCamelCase_ = chkpt['params'] lowerCamelCase_ = {n: v for n, v in config.items() if not isinstance(lowerCamelCase__ , (torch.FloatTensor, numpy.ndarray) )} lowerCamelCase_ = chkpt['dico_word2id'] lowerCamelCase_ = {s + '</w>' if s.find("@@" ) == -1 and i > 1_3 else s.replace("@@" , "" ): i for s, i in vocab.items()} # Save pytorch-model lowerCamelCase_ = pytorch_dump_folder_path + '/' + WEIGHTS_NAME lowerCamelCase_ = pytorch_dump_folder_path + '/' + CONFIG_NAME lowerCamelCase_ = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(lowerCamelCase__ , lowerCamelCase__ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(lowerCamelCase__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowerCamelCase__ , indent=2 ) + "\n" ) print(F'Save vocab file to {pytorch_config_dump_path}' ) with open(lowerCamelCase__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowerCamelCase__ , indent=2 ) + "\n" ) if __name__ == "__main__": __A =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __A =parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
463
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
0
"""simple docstring""" from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase ( __snake_case ): lowercase : str = ["image_processor", "tokenizer"] lowercase : Union[str, Any] = "AutoImageProcessor" lowercase : Union[str, Any] = "AutoTokenizer" def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : Union[str, Any] = self.image_processor def __call__( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCamelCase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: UpperCamelCase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: UpperCamelCase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def a_ ( self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def a_ ( self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def a_ ( self ): return ["input_ids", "attention_mask", "pixel_values"]
499
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
0
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging lowercase = logging.get_logger(__name__) lowercase = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class __lowerCamelCase ( __snake_case ): '''simple docstring''' snake_case__ : Optional[int] = "van" def __init__( self , a__=224 , a__=3 , a__=[7, 3, 3, 3] , a__=[4, 2, 2, 2] , a__=[64, 128, 320, 512] , a__=[3, 3, 12, 3] , a__=[8, 8, 4, 4] , a__="gelu" , a__=0.02 , a__=1e-6 , a__=1e-2 , a__=0.0 , a__=0.0 , **a__ , ): super().__init__(**_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Tuple = image_size __SCREAMING_SNAKE_CASE : Dict = num_channels __SCREAMING_SNAKE_CASE : Any = patch_sizes __SCREAMING_SNAKE_CASE : List[str] = strides __SCREAMING_SNAKE_CASE : List[Any] = hidden_sizes __SCREAMING_SNAKE_CASE : Optional[int] = depths __SCREAMING_SNAKE_CASE : Dict = mlp_ratios __SCREAMING_SNAKE_CASE : Dict = hidden_act __SCREAMING_SNAKE_CASE : List[str] = initializer_range __SCREAMING_SNAKE_CASE : Optional[Any] = layer_norm_eps __SCREAMING_SNAKE_CASE : Optional[int] = layer_scale_init_value __SCREAMING_SNAKE_CASE : Tuple = drop_path_rate __SCREAMING_SNAKE_CASE : Any = dropout_rate
211
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
0
'''simple docstring''' import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __magic_name__ : def __init__( self , snake_case_ , snake_case_=2 , snake_case_=3 , snake_case_=4 , snake_case_=2 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=36 , snake_case_=3 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=6 , snake_case_=6 , snake_case_=3 , snake_case_=4 , snake_case_=None , snake_case_=10_00 , ): lowercase =parent lowercase =batch_size lowercase =num_channels lowercase =image_size lowercase =patch_size lowercase =text_seq_length lowercase =is_training lowercase =use_input_mask lowercase =use_token_type_ids lowercase =use_labels lowercase =vocab_size lowercase =hidden_size lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =intermediate_size lowercase =hidden_act lowercase =hidden_dropout_prob lowercase =attention_probs_dropout_prob lowercase =max_position_embeddings lowercase =type_vocab_size lowercase =type_sequence_label_size lowercase =initializer_range lowercase =coordinate_size lowercase =shape_size lowercase =num_labels lowercase =num_choices lowercase =scope lowercase =range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) lowercase =text_seq_length lowercase =(image_size // patch_size) ** 2 + 1 lowercase =self.text_seq_length + self.image_seq_length def _A( self ): lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) lowercase =ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowercase =bbox[i, j, 3] lowercase =bbox[i, j, 1] lowercase =t if bbox[i, j, 2] < bbox[i, j, 0]: lowercase =bbox[i, j, 2] lowercase =bbox[i, j, 0] lowercase =t lowercase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase =None if self.use_input_mask: lowercase =random_attention_mask([self.batch_size, self.text_seq_length] ) lowercase =None if self.use_token_type_ids: lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) lowercase =None lowercase =None if self.use_labels: lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) lowercase =LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =LayoutLMvaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() # text + image lowercase =model(_lowerCAmelCase , pixel_values=_lowerCAmelCase ) lowercase =model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) lowercase =model(_lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) lowercase =model(_lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only lowercase =model(_lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only lowercase =model(pixel_values=_lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =self.num_labels lowercase =LayoutLMvaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowercase =model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =self.num_labels lowercase =LayoutLMvaForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowercase =model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowercase =LayoutLMvaForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowercase =model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _A( self ): lowercase =self.prepare_config_and_inputs() ( lowercase ) =config_and_inputs lowercase ={ 'input_ids': input_ids, 'bbox': bbox, 'pixel_values': pixel_values, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase__ = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def _A( self ): lowercase =LayoutLMvaModelTester(self ) lowercase =ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def _A( self , snake_case_ , snake_case_ , snake_case_=False ): lowercase =copy.deepcopy(_lowerCAmelCase ) if model_class in get_values(_lowerCAmelCase ): lowercase ={ k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(_lowerCAmelCase , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(_lowerCAmelCase ): lowercase =torch.ones(self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in get_values(_lowerCAmelCase ): lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in [ *get_values(_lowerCAmelCase ), ]: lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in [ *get_values(_lowerCAmelCase ), ]: lowercase =torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=_lowerCAmelCase , ) return inputs_dict def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase =type self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) @slow def _A( self ): for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =LayoutLMvaModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def UpperCamelCase ( ) -> Any: '''simple docstring''' lowercase =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch class __magic_name__ ( unittest.TestCase ): @cached_property def _A( self ): return LayoutLMvaImageProcessor(apply_ocr=_lowerCAmelCase ) if is_vision_available() else None @slow def _A( self ): lowercase =LayoutLMvaModel.from_pretrained('''microsoft/layoutlmv3-base''' ).to(_lowerCAmelCase ) lowercase =self.default_image_processor lowercase =prepare_img() lowercase =image_processor(images=_lowerCAmelCase , return_tensors='''pt''' ).pixel_values.to(_lowerCAmelCase ) lowercase =torch.tensor([[1, 2]] ) lowercase =torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass lowercase =model( input_ids=input_ids.to(_lowerCAmelCase ) , bbox=bbox.to(_lowerCAmelCase ) , pixel_values=pixel_values.to(_lowerCAmelCase ) , ) # verify the logits lowercase =torch.Size((1, 1_99, 7_68) ) self.assertEqual(outputs.last_hidden_state.shape , _lowerCAmelCase ) lowercase =torch.tensor( [[-0.05_29, 0.36_18, 0.16_32], [-0.15_87, -0.16_67, -0.04_00], [-0.15_57, -0.16_71, -0.05_05]] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , _lowerCAmelCase , atol=1E-4 ) )
72
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
0
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class A ( __snake_case , unittest.TestCase ): __UpperCAmelCase : List[str] = BarthezTokenizer __UpperCAmelCase : List[str] = BarthezTokenizerFast __UpperCAmelCase : List[Any] = True __UpperCAmelCase : Tuple = True def lowercase_ (self : Any ) -> List[Any]: """simple docstring""" super().setUp() UpperCAmelCase__ = BarthezTokenizerFast.from_pretrained("moussaKam/mbarthez" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_lowerCAmelCase ) UpperCAmelCase__ = tokenizer def lowercase_ (self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = '<pad>' UpperCAmelCase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def lowercase_ (self : Optional[int] ) -> str: """simple docstring""" UpperCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(_lowerCAmelCase ) , 1_0_1_1_2_2 ) def lowercase_ (self : Optional[int] ) -> str: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def lowercase_ (self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] UpperCAmelCase__ = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] UpperCAmelCase__ = self.tokenizer( _lowerCAmelCase , max_length=len(_lowerCAmelCase ) , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="pt" ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) UpperCAmelCase__ = batch.input_ids.tolist()[0] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def lowercase_ (self : List[Any] ) -> Union[str, Any]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase__ = self.get_tokenizer() UpperCAmelCase__ = self.get_rust_tokenizer() UpperCAmelCase__ = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ = self.get_rust_tokenizer() UpperCAmelCase__ = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) @slow def lowercase_ (self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = {'input_ids': [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. UpperCAmelCase__ = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name="moussaKam/mbarthez" , revision="c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6" , sequences=_lowerCAmelCase , )
486
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class snake_case__ (unittest.TestCase ): """simple docstring""" def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , __lowercase=True , ) -> Optional[int]: """simple docstring""" a__ : Optional[Any] = size if size is not None else {'shortest_edge': 2_0} a__ : List[str] = crop_size if crop_size is not None else {'height': 1_8, 'width': 1_8} a__ : Optional[Any] = parent a__ : List[Any] = batch_size a__ : Tuple = num_channels a__ : Optional[Any] = image_size a__ : str = min_resolution a__ : Optional[Any] = max_resolution a__ : Union[str, Any] = do_resize a__ : List[Any] = size a__ : List[str] = do_center_crop a__ : Tuple = crop_size a__ : List[Any] = do_flip_channel_order def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class snake_case__ (__snake_case , unittest.TestCase ): """simple docstring""" __lowerCAmelCase :Optional[int] = MobileViTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" a__ : List[str] = MobileViTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" a__ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_flip_channel_order""" ) ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 2_0} ) self.assertEqual(image_processor.crop_size , {"""height""": 1_8, """width""": 1_8} ) a__ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 4_2} ) self.assertEqual(image_processor.crop_size , {"""height""": 8_4, """width""": 8_4} ) def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" pass def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" a__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a__ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , Image.Image ) # Test not batched input a__ : Dict = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched a__ : List[str] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors a__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , numpify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , np.ndarray ) # Test not batched input a__ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched a__ : Union[str, Any] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" a__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors a__ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , torchify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , torch.Tensor ) # Test not batched input a__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched a__ : List[Any] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
136
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
0
import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort lowercase : Dict = logging.get_logger(__name__) lowercase : List[Any] = { "tensor(bool)": np.bool_, "tensor(int8)": np.inta, "tensor(uint8)": np.uinta, "tensor(int16)": np.intaa, "tensor(uint16)": np.uintaa, "tensor(int32)": np.intaa, "tensor(uint32)": np.uintaa, "tensor(int64)": np.intaa, "tensor(uint64)": np.uintaa, "tensor(float16)": np.floataa, "tensor(float)": np.floataa, "tensor(double)": np.floataa, } class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , __UpperCamelCase=None , **__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) __UpperCamelCase : Optional[Any] = model __UpperCamelCase : Optional[Any] = kwargs.get("model_save_dir" , _lowerCAmelCase ) __UpperCamelCase : Optional[Any] = kwargs.get("latest_model_name" , _lowerCAmelCase ) def __call__( self , **__UpperCamelCase ) -> int: '''simple docstring''' __UpperCamelCase : int = {k: np.array(_lowerCAmelCase ) for k, v in kwargs.items()} return self.model.run(_lowerCAmelCase , _lowerCAmelCase ) @staticmethod def __lowerCamelCase ( __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None ) -> Optional[int]: '''simple docstring''' if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) __UpperCamelCase : Any = 'CPUExecutionProvider' return ort.InferenceSession(_lowerCAmelCase , providers=[provider] , sess_options=_lowerCAmelCase ) def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : Any = file_name if file_name is not None else ONNX_WEIGHTS_NAME __UpperCamelCase : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) __UpperCamelCase : Union[str, Any] = Path(_lowerCAmelCase ).joinpath(_lowerCAmelCase ) try: shutil.copyfile(_lowerCAmelCase , _lowerCAmelCase ) except shutil.SameFileError: pass # copy external weights (for models >2GB) __UpperCamelCase : Dict = self.model_save_dir.joinpath(_lowerCAmelCase ) if src_path.exists(): __UpperCamelCase : str = Path(_lowerCAmelCase ).joinpath(_lowerCAmelCase ) try: shutil.copyfile(_lowerCAmelCase , _lowerCAmelCase ) except shutil.SameFileError: pass def __lowerCamelCase ( self , __UpperCamelCase , **__UpperCamelCase , ) -> int: '''simple docstring''' if os.path.isfile(_lowerCAmelCase ): logger.error(f'''Provided path ({save_directory}) should be a directory, not a file''' ) return os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) # saving model weights/files self._save_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) @classmethod def __lowerCamelCase ( cls , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = False , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , **__UpperCamelCase , ) -> List[Any]: '''simple docstring''' __UpperCamelCase : Optional[Any] = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(_lowerCAmelCase ): __UpperCamelCase : Optional[int] = OnnxRuntimeModel.load_model( os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , provider=_lowerCAmelCase , sess_options=_lowerCAmelCase ) __UpperCamelCase : Optional[int] = Path(_lowerCAmelCase ) # load model from hub else: # download model __UpperCamelCase : Any = hf_hub_download( repo_id=_lowerCAmelCase , filename=_lowerCAmelCase , use_auth_token=_lowerCAmelCase , revision=_lowerCAmelCase , cache_dir=_lowerCAmelCase , force_download=_lowerCAmelCase , ) __UpperCamelCase : Any = Path(_lowerCAmelCase ).parent __UpperCamelCase : str = Path(_lowerCAmelCase ).name __UpperCamelCase : Tuple = OnnxRuntimeModel.load_model(_lowerCAmelCase , provider=_lowerCAmelCase , sess_options=_lowerCAmelCase ) return cls(model=_lowerCAmelCase , **_lowerCAmelCase ) @classmethod def __lowerCamelCase ( cls , __UpperCamelCase , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = None , **__UpperCamelCase , ) -> Dict: '''simple docstring''' __UpperCamelCase : List[str] = None if len(str(_lowerCAmelCase ).split("@" ) ) == 2: __UpperCamelCase : List[str] = model_id.split("@" ) return cls._from_pretrained( model_id=_lowerCAmelCase , revision=_lowerCAmelCase , cache_dir=_lowerCAmelCase , force_download=_lowerCAmelCase , use_auth_token=_lowerCAmelCase , **_lowerCAmelCase , )
327
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
0
'''simple docstring''' import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() lowerCamelCase :List[Any] = logging.get_logger(__name__) lowerCamelCase :Optional[Any] = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } lowerCamelCase :Optional[int] = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' for attribute in key.split(""".""" ): A_ : Union[str, Any] = getattr(lowerCamelCase__ , lowerCamelCase__ ) if weight_type is not None: A_ : List[str] = getattr(lowerCamelCase__ , lowerCamelCase__ ).shape else: A_ : int = hf_pointer.shape assert hf_shape == value.shape, ( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": A_ : List[Any] = value elif weight_type == "weight_g": A_ : Optional[Any] = value elif weight_type == "weight_v": A_ : Any = value elif weight_type == "bias": A_ : Any = value else: A_ : int = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = [] A_ : int = fairseq_model.state_dict() A_ : str = hf_model.feature_extractor A_ : Tuple = hf_model.adapter for name, value in fairseq_dict.items(): A_ : Any = False if "conv_layers" in name: load_conv_layer( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , hf_model.config.feat_extract_norm == """group""" , ) A_ : Tuple = True elif any(x in name for x in ["""adaptor""", """w2v_encoder.proj.""", """w2v_proj_ln."""] ): load_adapter(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) A_ : Tuple = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: A_ : int = True if "*" in mapped_key: A_ : Tuple = name.split(lowerCamelCase__ )[0].split(""".""" )[-2] A_ : Optional[int] = mapped_key.replace("""*""" , lowerCamelCase__ ) if "weight_g" in name: A_ : Tuple = 'weight_g' elif "weight_v" in name: A_ : List[str] = 'weight_v' elif "bias" in name: A_ : Optional[int] = 'bias' elif "weight" in name: A_ : Tuple = 'weight' else: A_ : int = None set_recursively(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) continue if not is_used: unused_weights.append(lowerCamelCase__ ) logger.warning(f'Unused weights: {unused_weights}' ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : int = full_name.split("""conv_layers.""" )[-1] A_ : Dict = name.split(""".""" ) A_ : List[Any] = int(items[0] ) A_ : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) A_ : Any = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) A_ : int = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) A_ : Tuple = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) A_ : Union[str, Any] = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(lowerCamelCase__ ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Dict = full_name.split("""adaptor.""" )[-1] A_ : Optional[int] = name.split(""".""" ) if items[1].isdigit(): A_ : Any = int(items[1] ) else: A_ : Any = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), f'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.' A_ : Any = value logger.info(f'Adapter proj layer norm bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), f'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.' A_ : str = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), f'{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.' A_ : int = value logger.info(f'Adapter proj layer bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), f'{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.' A_ : Optional[Any] = value logger.info(f'Adapter proj layer weight was initialized from {full_name}.' ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), f'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.' A_ : Union[str, Any] = value logger.info(f'Adapter layer {layer_id} bias was initialized from {full_name}.' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), f'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.' A_ : Optional[int] = value logger.info(f'Adapter layer {layer_id} bias was initialized from {full_name}.' ) else: unused_weights.append(lowerCamelCase__ ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Tuple = emb.weight.shape A_ : Optional[int] = nn.Linear(lowerCamelCase__ , lowerCamelCase__ , bias=lowerCamelCase__ ) A_ : int = emb.weight.data return lin_layer @torch.no_grad() def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): '''simple docstring''' A_ : List[str] = WavaVecaConfig.from_pretrained( lowerCamelCase__ , add_adapter=lowerCamelCase__ , adapter_stride=lowerCamelCase__ , adapter_kernel_size=lowerCamelCase__ , use_auth_token=lowerCamelCase__ , output_hidden_size=lowerCamelCase__ , ) A_ : Optional[Any] = MBartConfig.from_pretrained(lowerCamelCase__ ) # load model A_ : Any = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ """config_yaml""": config_yaml_path, """data""": """/""".join(dict_path.split("""/""" )[:-1] ), """w2v_path""": checkpoint_path, """load_pretrained_decoder_from""": None, } , ) A_ : int = model[0].eval() # load feature extractor A_ : int = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase__ , use_auth_token=lowerCamelCase__ ) # set weights for wav2vec2 encoder A_ : str = WavaVecaModel(lowerCamelCase__ ) recursively_load_weights_wavaveca(model.encoder , lowerCamelCase__ ) # load decoder weights A_ : Any = MBartForCausalLM(lowerCamelCase__ ) A_ : Dict = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=lowerCamelCase__ ) logger.warning(f'The following keys are missing when loading the decoder weights: {missing_keys}' ) logger.warning(f'The following keys are unexpected when loading the decoder weights: {unexpected_keys}' ) A_ : List[Any] = SpeechEncoderDecoderModel(encoder=lowerCamelCase__ , decoder=lowerCamelCase__ ) A_ : int = False A_ : Tuple = MBartaaTokenizer(lowerCamelCase__ ) tokenizer.save_pretrained(lowerCamelCase__ ) A_ : List[str] = hf_wavavec.config.to_dict() A_ : List[str] = tokenizer.pad_token_id A_ : List[Any] = tokenizer.bos_token_id A_ : List[str] = tokenizer.eos_token_id A_ : int = 'mbart50' A_ : Tuple = 'wav2vec2' A_ : Optional[Any] = tokenizer.eos_token_id A_ : Optional[Any] = 25_00_04 A_ : Dict = tokenizer.eos_token_id A_ : int = SpeechEncoderDecoderConfig.from_dict(lowerCamelCase__ ) hf_wavavec.save_pretrained(lowerCamelCase__ ) feature_extractor.save_pretrained(lowerCamelCase__ ) if __name__ == "__main__": lowerCamelCase :List[Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_yaml_path''', default=None, type=str, help='''Path to yaml file of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-xls-r-1b''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/mbart-large-50-one-to-many-mmt''', type=str, help='''Path to hf decoder checkpoint config''', ) parser.add_argument('''--add_adapter''', default=True, type=bool, help='''whethere to add model adapter layers''') parser.add_argument('''--adapter_stride''', default=2, type=int, help='''stride of adapter layers''') parser.add_argument('''--adapter_kernel_size''', default=3, type=int, help='''kernel size of adapter layers''') parser.add_argument('''--encoder_output_dim''', default=1_0_2_4, type=int, help='''encoder output dim''') parser.add_argument('''--start_token_id''', default=2_5_0_0_0_4, type=int, help='''`decoder_start_token_id` of model config''') lowerCamelCase :List[str] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
667
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __A : Dict = logging.get_logger(__name__) __A : List[Any] = { '''microsoft/swin-tiny-patch4-window7-224''': ( '''https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json''' ), # See all Swin models at https://huggingface.co/models?filter=swin } class __A ( __snake_case , __snake_case ): lowerCAmelCase_ : List[Any] = "swin" lowerCAmelCase_ : Tuple = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : Optional[Any] , UpperCAmelCase_ : int=224 , UpperCAmelCase_ : int=4 , UpperCAmelCase_ : Tuple=3 , UpperCAmelCase_ : Optional[int]=96 , UpperCAmelCase_ : int=[2, 2, 6, 2] , UpperCAmelCase_ : Any=[3, 6, 12, 24] , UpperCAmelCase_ : Union[str, Any]=7 , UpperCAmelCase_ : str=4.0 , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : Optional[int]=0.0 , UpperCAmelCase_ : Tuple=0.0 , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : Union[str, Any]="gelu" , UpperCAmelCase_ : Tuple=False , UpperCAmelCase_ : int=0.02 , UpperCAmelCase_ : Dict=1E-5 , UpperCAmelCase_ : Optional[int]=32 , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : Optional[Any]=None , **UpperCAmelCase_ : Optional[int] , ): super().__init__(**_lowerCAmelCase ) lowerCAmelCase : Dict = image_size lowerCAmelCase : Optional[int] = patch_size lowerCAmelCase : str = num_channels lowerCAmelCase : List[Any] = embed_dim lowerCAmelCase : Optional[int] = depths lowerCAmelCase : Any = len(_lowerCAmelCase ) lowerCAmelCase : Tuple = num_heads lowerCAmelCase : List[Any] = window_size lowerCAmelCase : int = mlp_ratio lowerCAmelCase : Any = qkv_bias lowerCAmelCase : Any = hidden_dropout_prob lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase : List[str] = drop_path_rate lowerCAmelCase : Dict = hidden_act lowerCAmelCase : Tuple = use_absolute_embeddings lowerCAmelCase : Any = layer_norm_eps lowerCAmelCase : Dict = initializer_range lowerCAmelCase : Dict = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase : int = int(embed_dim * 2 ** (len(_lowerCAmelCase ) - 1) ) lowerCAmelCase : str = ['stem'] + [f"stage{idx}" for idx in range(1 , len(_lowerCAmelCase ) + 1 )] lowerCAmelCase : Any = get_aligned_output_features_output_indices( out_features=_lowerCAmelCase , out_indices=_lowerCAmelCase , stage_names=self.stage_names ) class __A ( __snake_case ): lowerCAmelCase_ : str = version.parse("1.11" ) @property def lowercase__ ( self : Any ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase__ ( self : Any ): return 1E-4
343
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = { 'configuration_mobilevit': ['MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileViTConfig', 'MobileViTOnnxConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['MobileViTFeatureExtractor'] __lowerCAmelCase = ['MobileViTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileViTForImageClassification', 'MobileViTForSemanticSegmentation', 'MobileViTModel', 'MobileViTPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFMobileViTForImageClassification', 'TFMobileViTForSemanticSegmentation', 'TFMobileViTModel', 'TFMobileViTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
201
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
0
'''simple docstring''' from __future__ import annotations import requests def UpperCAmelCase_ ( __lowercase : Tuple ) -> dict: '''simple docstring''' _UpperCAmelCase = f'https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty' return requests.get(__lowercase ).json() def UpperCAmelCase_ ( __lowercase : Optional[int] = 10 ) -> list[dict]: '''simple docstring''' _UpperCAmelCase = 'https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty' _UpperCAmelCase = requests.get(__lowercase ).json()[:max_stories] return [get_hackernews_story(__lowercase ) for story_id in story_ids] def UpperCAmelCase_ ( __lowercase : Tuple = 10 ) -> str: '''simple docstring''' _UpperCAmelCase = hackernews_top_stories(__lowercase ) return "\n".join("* [{title}]({url})".format(**__lowercase ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
236
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
0
from math import isqrt def lowerCamelCase_ ( lowerCamelCase__ ): return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCamelCase__ ) + 1 ) ) def lowerCamelCase_ ( lowerCamelCase__ = 1_0**6 ): lowerCamelCase_ = 0 lowerCamelCase_ = 1 lowerCamelCase_ = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCamelCase__ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
463
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class lowerCamelCase ( unittest.TestCase ): def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=13 , SCREAMING_SNAKE_CASE_=7 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=99 , SCREAMING_SNAKE_CASE_=32 , SCREAMING_SNAKE_CASE_=5 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=37 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=512 , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=4 , ): UpperCamelCase : Optional[Any] = parent UpperCamelCase : Any = batch_size UpperCamelCase : str = seq_length UpperCamelCase : Union[str, Any] = is_training UpperCamelCase : Tuple = use_attention_mask UpperCamelCase : List[str] = use_token_type_ids UpperCamelCase : Tuple = use_labels UpperCamelCase : Tuple = vocab_size UpperCamelCase : List[Any] = hidden_size UpperCamelCase : List[Any] = num_hidden_layers UpperCamelCase : List[Any] = num_attention_heads UpperCamelCase : Any = intermediate_size UpperCamelCase : List[str] = hidden_act UpperCamelCase : Dict = hidden_dropout_prob UpperCamelCase : int = attention_probs_dropout_prob UpperCamelCase : str = max_position_embeddings UpperCamelCase : Tuple = type_vocab_size UpperCamelCase : Optional[int] = type_sequence_label_size UpperCamelCase : Optional[int] = initializer_range UpperCamelCase : List[Any] = num_choices def a_ ( self ): UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_attention_mask: UpperCamelCase : int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Optional[Any] = None if self.use_token_type_ids: UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Any = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def a_ ( self ): UpperCamelCase : List[Any] = self.prepare_config_and_inputs() UpperCamelCase : Dict = config_and_inputs UpperCamelCase : Dict = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def a_ ( self ): UpperCamelCase : Tuple = self.prepare_config_and_inputs() UpperCamelCase : List[Any] = config_and_inputs UpperCamelCase : Any = True UpperCamelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowerCamelCase ( __snake_case , unittest.TestCase ): lowercase : Optional[Any] = True lowercase : int = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def a_ ( self ): UpperCamelCase : Optional[Any] = FlaxBertModelTester(self ) @slow def a_ ( self ): # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. UpperCamelCase : Union[str, Any] = FlaxBertModel.from_pretrained("""bert-base-cased""" ) UpperCamelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowerCAmelCase )
499
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
0
'''simple docstring''' import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class __lowerCamelCase ( __snake_case , unittest.TestCase ): '''simple docstring''' snake_case__ : List[Any] = MvpTokenizer snake_case__ : Optional[Any] = MvpTokenizerFast snake_case__ : List[Any] = True snake_case__ : Dict = filter_roberta_detectors def a_ ( self ): super().setUp() __SCREAMING_SNAKE_CASE : Optional[int] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __SCREAMING_SNAKE_CASE : str = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __SCREAMING_SNAKE_CASE : Optional[int] = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __SCREAMING_SNAKE_CASE : Optional[Any] = {'unk_token': '<unk>'} __SCREAMING_SNAKE_CASE : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __SCREAMING_SNAKE_CASE : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_lowerCAmelCase ) ) def a_ ( self , **a__ ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def a_ ( self , **a__ ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def a_ ( self , a__ ): return "lower newer", "lower newer" @cached_property def a_ ( self ): return MvpTokenizer.from_pretrained("RUCAIBox/mvp" ) @cached_property def a_ ( self ): return MvpTokenizerFast.from_pretrained("RUCAIBox/mvp" ) @require_torch def a_ ( self ): __SCREAMING_SNAKE_CASE : int = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] __SCREAMING_SNAKE_CASE : Tuple = [0, 250, 251, 17818, 13, 39186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __SCREAMING_SNAKE_CASE : List[Any] = tokenizer(_lowerCAmelCase , max_length=len(_lowerCAmelCase ) , padding=_lowerCAmelCase , return_tensors="pt" ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) __SCREAMING_SNAKE_CASE : str = batch.input_ids.tolist()[0] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # Test that special tokens are reset @require_torch def a_ ( self ): __SCREAMING_SNAKE_CASE : Dict = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __SCREAMING_SNAKE_CASE : str = tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors="pt" ) # check if input_ids are returned and no labels self.assertIn("input_ids" , _lowerCAmelCase ) self.assertIn("attention_mask" , _lowerCAmelCase ) self.assertNotIn("labels" , _lowerCAmelCase ) self.assertNotIn("decoder_attention_mask" , _lowerCAmelCase ) @require_torch def a_ ( self ): __SCREAMING_SNAKE_CASE : Optional[int] = [ 'Summary of the text.', 'Another summary.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __SCREAMING_SNAKE_CASE : List[str] = tokenizer(text_target=_lowerCAmelCase , max_length=32 , padding="max_length" , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) @require_torch def a_ ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer( ["I am a small frog" * 1024, "I am a small frog"] , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="pt" ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(batch.input_ids.shape , (2, 1024) ) @require_torch def a_ ( self ): __SCREAMING_SNAKE_CASE : Tuple = ['A long paragraph for summarization.'] __SCREAMING_SNAKE_CASE : Any = [ 'Summary of the text.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer(_lowerCAmelCase , text_target=_lowerCAmelCase , return_tensors="pt" ) __SCREAMING_SNAKE_CASE : Any = inputs['input_ids'] __SCREAMING_SNAKE_CASE : Union[str, Any] = inputs['labels'] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def a_ ( self ): pass def a_ ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): __SCREAMING_SNAKE_CASE : Optional[int] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Optional[Any] = 'A, <mask> AllenNLP sentence.' __SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer_r.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer_p.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) __SCREAMING_SNAKE_CASE : Dict = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) __SCREAMING_SNAKE_CASE : Dict = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( _lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( _lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
211
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
0
'''simple docstring''' import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[int] = '''▁''' _UpperCAmelCase : List[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class __magic_name__ ( __snake_case , unittest.TestCase ): UpperCamelCase__ = BertGenerationTokenizer UpperCamelCase__ = False UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BertGenerationTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _A( self ): lowercase ='<s>' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''<pad>''' ) self.assertEqual(len(_lowerCAmelCase ) , 10_02 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def _A( self ): lowercase =BertGenerationTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) lowercase =tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowerCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [2_85, 46, 10, 1_70, 3_82] , ) lowercase =tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowercase =tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowercase =tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def _A( self ): return BertGenerationTokenizer.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) @slow def _A( self ): lowercase ='Hello World!' lowercase =[1_85_36, 22_60, 1_01] self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase ) ) @slow def _A( self ): lowercase =( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowercase =[ 8_71, 4_19, 3_58, 9_46, 9_91, 25_21, 4_52, 3_58, 13_57, 3_87, 77_51, 35_36, 1_12, 9_85, 4_56, 1_26, 8_65, 9_38, 54_00, 57_34, 4_58, 13_68, 4_67, 7_86, 24_62, 52_46, 11_59, 6_33, 8_65, 45_19, 4_57, 5_82, 8_52, 25_57, 4_27, 9_16, 5_08, 4_05, 3_43_24, 4_97, 3_91, 4_08, 1_13_42, 12_44, 3_85, 1_00, 9_38, 9_85, 4_56, 5_74, 3_62, 1_25_97, 32_00, 31_29, 11_72, ] self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase ) ) @require_torch @slow def _A( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowercase =list(self.big_tokenizer.get_vocab().keys() )[:10] lowercase =' '.join(_lowerCAmelCase ) lowercase =self.big_tokenizer.encode_plus(_lowerCAmelCase , return_tensors='''pt''' , return_token_type_ids=_lowerCAmelCase ) lowercase =self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=_lowerCAmelCase ) lowercase =BertGenerationConfig() lowercase =BertGenerationEncoder(_lowerCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_lowerCAmelCase ) model(**_lowerCAmelCase ) @slow def _A( self ): # fmt: off lowercase ={'input_ids': [[3_92_86, 4_58, 3_63_35, 20_01, 4_56, 1_30_73, 1_32_66, 4_55, 1_13, 77_46, 17_41, 1_11_57, 3_91, 1_30_73, 1_32_66, 4_55, 1_13, 39_67, 3_54_12, 1_13, 49_36, 1_09, 38_70, 23_77, 1_13, 3_00_84, 4_57_20, 4_58, 1_34, 1_74_96, 1_12, 5_03, 1_16_72, 1_13, 1_18, 1_12, 56_65, 1_33_47, 3_86_87, 1_12, 14_96, 3_13_89, 1_12, 32_68, 4_72_64, 1_34, 9_62, 1_12, 1_63_77, 80_35, 2_31_30, 4_30, 1_21_69, 1_55_18, 2_85_92, 4_58, 1_46, 4_16_97, 1_09, 3_91, 1_21_69, 1_55_18, 1_66_89, 4_58, 1_46, 4_13_58, 1_09, 4_52, 7_26, 40_34, 1_11, 7_63, 3_54_12, 50_82, 3_88, 19_03, 1_11, 90_51, 3_91, 28_70, 4_89_18, 19_00, 11_23, 5_50, 9_98, 1_12, 95_86, 1_59_85, 4_55, 3_91, 4_10, 2_29_55, 3_76_36, 1_14], [4_48, 1_74_96, 4_19, 36_63, 3_85, 7_63, 1_13, 2_75_33, 28_70, 32_83, 1_30_43, 16_39, 2_47_13, 5_23, 6_56, 2_40_13, 1_85_50, 25_21, 5_17, 2_70_14, 2_12_44, 4_20, 12_12, 14_65, 3_91, 9_27, 48_33, 3_88, 5_78, 1_17_86, 1_14, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_84, 21_69, 76_87, 2_19_32, 1_81_46, 7_26, 3_63, 1_70_32, 33_91, 1_14, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='''google/bert_for_seq_generation_L-24_bbc_encoder''' , revision='''c817d1fd1be2ffa69431227a1fe320544943d4db''' , )
72
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
0
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class A : __UpperCAmelCase : Dict = None def lowercase_ (self : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) UpperCAmelCase__ = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , _lowerCAmelCase ) def lowercase_ (self : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = os.path.join(_lowerCAmelCase , "feat_extract.json" ) feat_extract_first.to_json_file(_lowerCAmelCase ) UpperCAmelCase__ = self.feature_extraction_class.from_json_file(_lowerCAmelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = feat_extract_first.save_pretrained(_lowerCAmelCase )[0] check_json_file_has_correct_format(_lowerCAmelCase ) UpperCAmelCase__ = self.feature_extraction_class.from_pretrained(_lowerCAmelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def lowercase_ (self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.feature_extraction_class() self.assertIsNotNone(_lowerCAmelCase )
486
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class snake_case__ : """simple docstring""" def __init__( self ) -> List[Any]: """simple docstring""" a__ : Optional[Any] = '' a__ : List[Any] = '' a__ : Union[str, Any] = [] a__ : List[str] = 0 a__ : Dict = 2_5_6 a__ : Optional[Any] = 0 a__ : List[str] = 0 a__ : Union[str, Any] = 0 a__ : str = 0 def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[str]: """simple docstring""" a__ : Union[str, Any] = cva.imread(_lowerCAmelCase , 0 ) a__ : Tuple = copy.deepcopy(self.img ) a__ : Tuple = plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] , label="""x""" ) a__ : int = np.sum(_lowerCAmelCase ) for i in range(len(_lowerCAmelCase ) ): a__ : Dict = x[i] / self.k self.sk += prk a__ : int = (self.L - 1) * self.sk if self.rem != 0: a__ : Union[str, Any] = int(last % last ) a__ : int = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(_lowerCAmelCase ) a__ : List[Any] = int(np.ma.count(self.img ) / self.img[1].size ) a__ : Dict = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): a__ : Any = self.img[j][i] if num != self.last_list[num]: a__ : Union[str, Any] = self.last_list[num] cva.imwrite("""output_data/output.jpg""" , self.img ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" cva.imshow("""Output-Image""" , self.img ) cva.imshow("""Input-Image""" , self.original_image ) cva.waitKey(5_0_0_0 ) cva.destroyAllWindows() if __name__ == "__main__": _lowercase : Union[str, Any] =os.path.join(os.path.basename(__file__), "image_data/input.jpg") _lowercase : Dict =ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
136
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : str = logging.get_logger(__name__) lowercase : Dict = { "facebook/s2t-small-librispeech-asr": ( "https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class SCREAMING_SNAKE_CASE__ ( __snake_case ): """simple docstring""" lowercase : Union[str, Any] = "speech_to_text" lowercase : Optional[int] = ["past_key_values"] lowercase : List[Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self , __UpperCamelCase=1_00_00 , __UpperCamelCase=12 , __UpperCamelCase=20_48 , __UpperCamelCase=4 , __UpperCamelCase=6 , __UpperCamelCase=20_48 , __UpperCamelCase=4 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase="relu" , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.02 , __UpperCamelCase=2 , __UpperCamelCase=True , __UpperCamelCase=1 , __UpperCamelCase=0 , __UpperCamelCase=2 , __UpperCamelCase=60_00 , __UpperCamelCase=10_24 , __UpperCamelCase=2 , __UpperCamelCase=(5, 5) , __UpperCamelCase=10_24 , __UpperCamelCase=80 , __UpperCamelCase=1 , **__UpperCamelCase , ) -> Any: '''simple docstring''' __UpperCamelCase : Tuple = vocab_size __UpperCamelCase : int = d_model __UpperCamelCase : Tuple = encoder_ffn_dim __UpperCamelCase : Union[str, Any] = encoder_layers __UpperCamelCase : Dict = encoder_attention_heads __UpperCamelCase : int = decoder_ffn_dim __UpperCamelCase : Any = decoder_layers __UpperCamelCase : Optional[int] = decoder_attention_heads __UpperCamelCase : Optional[int] = dropout __UpperCamelCase : Optional[int] = attention_dropout __UpperCamelCase : str = activation_dropout __UpperCamelCase : int = activation_function __UpperCamelCase : Dict = init_std __UpperCamelCase : Any = encoder_layerdrop __UpperCamelCase : Optional[int] = decoder_layerdrop __UpperCamelCase : Tuple = use_cache __UpperCamelCase : int = encoder_layers __UpperCamelCase : List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True __UpperCamelCase : Any = max_source_positions __UpperCamelCase : Any = max_target_positions __UpperCamelCase : Any = num_conv_layers __UpperCamelCase : Union[str, Any] = list(_lowerCAmelCase ) __UpperCamelCase : List[Any] = conv_channels __UpperCamelCase : int = input_feat_per_channel __UpperCamelCase : Union[str, Any] = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` " f'''but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, ''' f'''`config.num_conv_layers = {self.num_conv_layers}`.''' ) super().__init__( pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , is_encoder_decoder=_lowerCAmelCase , decoder_start_token_id=_lowerCAmelCase , **_lowerCAmelCase , )
327
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
0
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=10_24 , lowerCamelCase__=10_24 , lowerCamelCase__=False , **lowerCamelCase__ ): '''simple docstring''' A_ : Optional[Any] = AutoTokenizer.from_pretrained(lowerCamelCase__ ) A_ : List[str] = SeqaSeqDataset(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , type_path="""train""" , **lowerCamelCase__ ) A_ : Dict = tok.pad_token_id def get_lens(lowerCamelCase__ ): A_ : Any = tqdm( DataLoader(lowerCamelCase__ , batch_size=5_12 , num_workers=8 , shuffle=lowerCamelCase__ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) A_ : Union[str, Any] = [] for batch in dl: A_ : Dict = batch['input_ids'].ne(lowerCamelCase__ ).sum(1 ).tolist() A_ : Union[str, Any] = batch['labels'].ne(lowerCamelCase__ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCamelCase__ , lowerCamelCase__ ): max_lens.append(max(lowerCamelCase__ , lowerCamelCase__ ) ) else: max_lens.extend(lowerCamelCase__ ) return max_lens A_ : str = get_lens(lowerCamelCase__ ) A_ : Optional[int] = SeqaSeqDataset(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , type_path="""val""" , **lowerCamelCase__ ) A_ : List[Any] = get_lens(lowerCamelCase__ ) pickle_save(lowerCamelCase__ , train_ds.len_file ) pickle_save(lowerCamelCase__ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
667
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING __A : Optional[int] = logging.get_logger(__name__) @add_end_docstrings(__snake_case ) class __A ( __snake_case ): def __init__( self : Optional[int] , **UpperCAmelCase_ : List[Any] ): super().__init__(**_lowerCAmelCase ) if self.framework == "tf": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) requires_backends(self , 'vision' ) self.check_model_type(_lowerCAmelCase ) def __call__( self : int , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[int] = None , **UpperCAmelCase_ : Any , ): if "text_queries" in kwargs: lowerCAmelCase : Union[str, Any] = kwargs.pop('text_queries' ) if isinstance(_lowerCAmelCase , (str, Image.Image) ): lowerCAmelCase : List[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowerCAmelCase : Union[str, Any] = image lowerCAmelCase : Dict = super().__call__(_lowerCAmelCase , **_lowerCAmelCase ) return results def lowercase__ ( self : List[str] , **UpperCAmelCase_ : int ): lowerCAmelCase : Optional[int] = {} if "threshold" in kwargs: lowerCAmelCase : Dict = kwargs['threshold'] if "top_k" in kwargs: lowerCAmelCase : int = kwargs['top_k'] return {}, {}, postprocess_params def lowercase__ ( self : str , UpperCAmelCase_ : Tuple ): lowerCAmelCase : Dict = load_image(inputs['image'] ) lowerCAmelCase : Any = inputs['candidate_labels'] if isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowerCAmelCase : Tuple = candidate_labels.split(',' ) lowerCAmelCase : int = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(_lowerCAmelCase ): lowerCAmelCase : Union[str, Any] = self.tokenizer(_lowerCAmelCase , return_tensors=self.framework ) lowerCAmelCase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(_lowerCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def lowercase__ ( self : int , UpperCAmelCase_ : List[str] ): lowerCAmelCase : Dict = model_inputs.pop('target_size' ) lowerCAmelCase : Optional[Any] = model_inputs.pop('candidate_label' ) lowerCAmelCase : int = model_inputs.pop('is_last' ) lowerCAmelCase : Dict = self.model(**_lowerCAmelCase ) lowerCAmelCase : Dict = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def lowercase__ ( self : Tuple , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple=0.1 , UpperCAmelCase_ : List[str]=None ): lowerCAmelCase : List[Any] = [] for model_output in model_outputs: lowerCAmelCase : List[str] = model_output['candidate_label'] lowerCAmelCase : Optional[int] = BaseModelOutput(_lowerCAmelCase ) lowerCAmelCase : List[str] = self.image_processor.post_process_object_detection( outputs=_lowerCAmelCase , threshold=_lowerCAmelCase , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowerCAmelCase : Tuple = outputs['scores'][index].item() lowerCAmelCase : Any = self._get_bounding_box(outputs['boxes'][index][0] ) lowerCAmelCase : Dict = {'score': score, 'label': label, 'box': box} results.append(_lowerCAmelCase ) lowerCAmelCase : List[str] = sorted(_lowerCAmelCase , key=lambda UpperCAmelCase_ : x["score"] , reverse=_lowerCAmelCase ) if top_k: lowerCAmelCase : List[Any] = results[:top_k] return results def lowercase__ ( self : List[Any] , UpperCAmelCase_ : List[str] ): if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowerCAmelCase : Optional[Any] = box.int().tolist() lowerCAmelCase : Dict = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
343
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCamelCase ( __snake_case , unittest.TestCase ): UpperCamelCase_ : List[str] = KandinskyImgaImgPipeline UpperCamelCase_ : List[str] = ["prompt", "image_embeds", "negative_image_embeds", "image"] UpperCamelCase_ : Dict = [ "prompt", "negative_prompt", "image_embeds", "negative_image_embeds", "image", ] UpperCamelCase_ : Optional[int] = [ "generator", "height", "width", "strength", "guidance_scale", "negative_prompt", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] UpperCamelCase_ : Tuple = False @property def snake_case__ ( self :Optional[int] ) -> Optional[Any]: """simple docstring""" return 3_2 @property def snake_case__ ( self :Tuple ) -> int: """simple docstring""" return 3_2 @property def snake_case__ ( self :Dict ) -> Optional[int]: """simple docstring""" return self.time_input_dim @property def snake_case__ ( self :int ) -> int: """simple docstring""" return self.time_input_dim * 4 @property def snake_case__ ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" return 1_0_0 @property def snake_case__ ( self :Optional[Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE = XLMRobertaTokenizerFast.from_pretrained('''YiYiXu/tiny-random-mclip-base''' ) return tokenizer @property def snake_case__ ( self :str ) -> Tuple: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=3_7 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1_0_0_5 , ) SCREAMING_SNAKE_CASE = MultilingualCLIP(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = text_encoder.eval() return text_encoder @property def snake_case__ ( self :Any ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } SCREAMING_SNAKE_CASE = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def snake_case__ ( self :str ) -> Tuple: """simple docstring""" return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def snake_case__ ( self :int ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = VQModel(**self.dummy_movq_kwargs ) return model def snake_case__ ( self :Optional[Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE = self.dummy_text_encoder SCREAMING_SNAKE_CASE = self.dummy_tokenizer SCREAMING_SNAKE_CASE = self.dummy_unet SCREAMING_SNAKE_CASE = self.dummy_movq SCREAMING_SNAKE_CASE = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } SCREAMING_SNAKE_CASE = DDIMScheduler(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def snake_case__ ( self :Union[str, Any] , lowercase :List[Any] , lowercase :str=0 ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(_lowerCAmelCase ) # create init_image SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = image.cpu().permute(0 , 2 , 3 , 1 )[0] SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert('''RGB''' ).resize((2_5_6, 2_5_6) ) if str(_lowerCAmelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE = torch.manual_seed(_lowerCAmelCase ) else: SCREAMING_SNAKE_CASE = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = { 'prompt': 'horse', 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def snake_case__ ( self :Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE = 'cpu' SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = self.pipeline_class(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) SCREAMING_SNAKE_CASE = output.images SCREAMING_SNAKE_CASE = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) SCREAMING_SNAKE_CASE = np.array( [0.61_47_49_43, 0.6_07_35_39, 0.43_30_85_44, 0.5_92_82_69, 0.47_49_35_95, 0.46_75_59_73, 0.4_61_38_38, 0.45_36_87_97, 0.50_11_92_33] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCamelCase ( unittest.TestCase ): def snake_case__ ( self :Dict ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self :Tuple ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/kandinsky_img2img_frog.npy''' ) SCREAMING_SNAKE_CASE = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) SCREAMING_SNAKE_CASE = 'A red cartoon frog, 4k' SCREAMING_SNAKE_CASE = KandinskyPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-1-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE = KandinskyImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-1''' , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE = torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() SCREAMING_SNAKE_CASE = pipeline( _lowerCAmelCase , image=_lowerCAmelCase , image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=1_0_0 , height=7_6_8 , width=7_6_8 , strength=0.2 , output_type='''np''' , ) SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
201
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
'''simple docstring''' def UpperCAmelCase_ ( __lowercase : List[Any] = 3 , __lowercase : Optional[Any] = 7 , __lowercase : Optional[int] = 100_0000 ) -> int: '''simple docstring''' _UpperCAmelCase = 0 _UpperCAmelCase = 1 for current_denominator in range(1 , limit + 1 ): _UpperCAmelCase = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: _UpperCAmelCase = current_numerator _UpperCAmelCase = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1000000))
236
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() __A =logging.get_logger(__name__) __A ={ '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''adapter_layer''': '''encoder.layers.*.adapter_layer''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', '''pooling_layer.linear''': '''projector''', '''pooling_layer.projection''': '''classifier''', } __A =[ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', '''projector''', '''classifier''', ] def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = {} with open(lowerCamelCase__ , "r" ) as file: for line_number, line in enumerate(lowerCamelCase__ ): lowerCamelCase_ = line.strip() if line: lowerCamelCase_ = line.split() lowerCamelCase_ = line_number lowerCamelCase_ = words[0] lowerCamelCase_ = value return result def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): for attribute in key.split("." ): lowerCamelCase_ = getattr(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCamelCase__ ): lowerCamelCase_ = PARAM_MAPPING[full_name.split("." )[-1]] lowerCamelCase_ = 'param' if weight_type is not None and weight_type != "param": lowerCamelCase_ = getattr(lowerCamelCase__ , lowerCamelCase__ ).shape elif weight_type is not None and weight_type == "param": lowerCamelCase_ = hf_pointer for attribute in hf_param_name.split("." ): lowerCamelCase_ = getattr(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = shape_pointer.shape # let's reduce dimension lowerCamelCase_ = value[0] else: lowerCamelCase_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": lowerCamelCase_ = value elif weight_type == "weight_g": lowerCamelCase_ = value elif weight_type == "weight_v": lowerCamelCase_ = value elif weight_type == "bias": lowerCamelCase_ = value elif weight_type == "param": for attribute in hf_param_name.split("." ): lowerCamelCase_ = getattr(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = value else: lowerCamelCase_ = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCamelCase__ ): lowerCamelCase_ = PARAM_MAPPING[full_name.split("." )[-1]] lowerCamelCase_ = 'param' if weight_type is not None and weight_type != "param": lowerCamelCase_ = '.'.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": lowerCamelCase_ = '.'.join([key, hf_param_name] ) else: lowerCamelCase_ = key lowerCamelCase_ = value if 'lm_head' in full_key else value[0] __A ={ '''W_a''': '''linear_1.weight''', '''W_b''': '''linear_2.weight''', '''b_a''': '''linear_1.bias''', '''b_b''': '''linear_2.bias''', '''ln_W''': '''norm.weight''', '''ln_b''': '''norm.bias''', } def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=None ): lowerCamelCase_ = False for key, mapped_key in MAPPING.items(): lowerCamelCase_ = 'wav2vec2.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: lowerCamelCase_ = True if "*" in mapped_key: lowerCamelCase_ = name.split(lowerCamelCase__ )[0].split("." )[-2] lowerCamelCase_ = mapped_key.replace("*" , lowerCamelCase__ ) if "weight_g" in name: lowerCamelCase_ = 'weight_g' elif "weight_v" in name: lowerCamelCase_ = 'weight_v' elif "bias" in name: lowerCamelCase_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj lowerCamelCase_ = 'weight' else: lowerCamelCase_ = None if hf_dict is not None: rename_dict(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) else: set_recursively(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return is_used return is_used def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = [] lowerCamelCase_ = fairseq_model.state_dict() lowerCamelCase_ = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): lowerCamelCase_ = False if "conv_layers" in name: load_conv_layer( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , hf_model.config.feat_extract_norm == "group" , ) lowerCamelCase_ = True else: lowerCamelCase_ = load_wavaveca_layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if not is_used: unused_weights.append(lowerCamelCase__ ) logger.warning(F'Unused weights: {unused_weights}' ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = full_name.split("conv_layers." )[-1] lowerCamelCase_ = name.split("." ) lowerCamelCase_ = int(items[0] ) lowerCamelCase_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(lowerCamelCase__ ) @torch.no_grad() def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=False ): if config_path is not None: lowerCamelCase_ = WavaVecaConfig.from_pretrained(lowerCamelCase__ ) else: lowerCamelCase_ = WavaVecaConfig() if is_seq_class: lowerCamelCase_ = read_txt_into_dict(lowerCamelCase__ ) lowerCamelCase_ = idalabel lowerCamelCase_ = WavaVecaForSequenceClassification(lowerCamelCase__ ) lowerCamelCase_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ) feature_extractor.save_pretrained(lowerCamelCase__ ) elif is_finetuned: if dict_path: lowerCamelCase_ = Dictionary.load(lowerCamelCase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCamelCase_ = target_dict.pad_index lowerCamelCase_ = target_dict.bos_index lowerCamelCase_ = target_dict.eos_index lowerCamelCase_ = len(target_dict.symbols ) lowerCamelCase_ = os.path.join(lowerCamelCase__ , "vocab.json" ) if not os.path.isdir(lowerCamelCase__ ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(lowerCamelCase__ ) ) return os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) lowerCamelCase_ = target_dict.indices # fairseq has the <pad> and <s> switched lowerCamelCase_ = 0 lowerCamelCase_ = 1 with open(lowerCamelCase__ , "w" , encoding="utf-8" ) as vocab_handle: json.dump(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = WavaVecaCTCTokenizer( lowerCamelCase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=lowerCamelCase__ , ) lowerCamelCase_ = True if config.feat_extract_norm == 'layer' else False lowerCamelCase_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ) lowerCamelCase_ = WavaVecaProcessor(feature_extractor=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) processor.save_pretrained(lowerCamelCase__ ) lowerCamelCase_ = WavaVecaForCTC(lowerCamelCase__ ) else: lowerCamelCase_ = WavaVecaForPreTraining(lowerCamelCase__ ) if is_finetuned or is_seq_class: lowerCamelCase_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) else: lowerCamelCase_ = argparse.Namespace(task="audio_pretraining" ) lowerCamelCase_ = fairseq.tasks.setup_task(lowerCamelCase__ ) lowerCamelCase_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowerCamelCase__ ) lowerCamelCase_ = model[0].eval() recursively_load_weights(lowerCamelCase__ , lowerCamelCase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowerCamelCase__ ) if __name__ == "__main__": __A =argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) parser.add_argument( '''--is_seq_class''', action='''store_true''', help='''Whether the model to convert is a fine-tuned sequence classification model or not''', ) __A =parser.parse_args() __A =not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
463
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
0
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowerCamelCase : def __init__( self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : int = data UpperCamelCase : Optional[int] = [0X67_45_23_01, 0Xef_cd_ab_89, 0X98_ba_dc_fe, 0X10_32_54_76, 0Xc3_d2_e1_f0] @staticmethod def a_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return ((n << b) | (n >> (32 - b))) & 0Xff_ff_ff_ff def a_ ( self ): UpperCamelCase : Union[str, Any] = b'\x80' + b'\x00' * (63 - (len(self.data ) + 8) % 64) UpperCamelCase : List[Any] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def a_ ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def a_ ( self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCamelCase : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def a_ ( self ): UpperCamelCase : Dict = self.padding() UpperCamelCase : int = self.split_blocks() for block in self.blocks: UpperCamelCase : List[str] = self.expand_block(_lowerCAmelCase ) UpperCamelCase : List[Any] = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCamelCase : Any = (b & c) | ((~b) & d) UpperCamelCase : List[Any] = 0X5a_82_79_99 elif 20 <= i < 40: UpperCamelCase : Optional[Any] = b ^ c ^ d UpperCamelCase : Union[str, Any] = 0X6e_d9_eb_a1 elif 40 <= i < 60: UpperCamelCase : Optional[int] = (b & c) | (b & d) | (c & d) UpperCamelCase : Tuple = 0X8f_1b_bc_dc elif 60 <= i < 80: UpperCamelCase : Any = b ^ c ^ d UpperCamelCase : Optional[Any] = 0Xca_62_c1_d6 UpperCamelCase : Union[str, Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xff_ff_ff_ff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCamelCase : Optional[Any] = ( self.h[0] + a & 0Xff_ff_ff_ff, self.h[1] + b & 0Xff_ff_ff_ff, self.h[2] + c & 0Xff_ff_ff_ff, self.h[3] + d & 0Xff_ff_ff_ff, self.h[4] + e & 0Xff_ff_ff_ff, ) return ("{:08x}" * 5).format(*self.h ) def A_ ( ): '''simple docstring''' UpperCamelCase : Union[str, Any] = b'Test String' assert SHAaHash(snake_case_ ).final_hash() == hashlib.shaa(snake_case_ ).hexdigest() # noqa: S324 def A_ ( ): '''simple docstring''' UpperCamelCase : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument("""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) UpperCamelCase : Optional[Any] = parser.parse_args() UpperCamelCase : Any = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: UpperCamelCase : Dict = f.read() else: UpperCamelCase : List[Any] = bytes(snake_case_ ,"""utf-8""" ) print(SHAaHash(snake_case_ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
499
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowerCamelCase ( __snake_case , unittest.TestCase ): '''simple docstring''' snake_case__ : Union[str, Any] = KandinskyVaaImgaImgPipeline snake_case__ : Optional[int] = ["image_embeds", "negative_image_embeds", "image"] snake_case__ : Any = [ "image_embeds", "negative_image_embeds", "image", ] snake_case__ : int = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] snake_case__ : str = False @property def a_ ( self ): return 32 @property def a_ ( self ): return 32 @property def a_ ( self ): return self.time_input_dim @property def a_ ( self ): return self.time_input_dim * 4 @property def a_ ( self ): return 100 @property def a_ ( self ): torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } __SCREAMING_SNAKE_CASE : str = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def a_ ( self ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def a_ ( self ): torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Optional[int] = VQModel(**self.dummy_movq_kwargs ) return model def a_ ( self ): __SCREAMING_SNAKE_CASE : Union[str, Any] = self.dummy_unet __SCREAMING_SNAKE_CASE : List[Any] = self.dummy_movq __SCREAMING_SNAKE_CASE : List[Any] = { 'num_train_timesteps': 1000, 'beta_schedule': 'linear', 'beta_start': 0.00085, 'beta_end': 0.012, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } __SCREAMING_SNAKE_CASE : str = DDIMScheduler(**_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def a_ ( self , a__ , a__=0 ): __SCREAMING_SNAKE_CASE : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : int = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowerCAmelCase ) # create init_image __SCREAMING_SNAKE_CASE : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] __SCREAMING_SNAKE_CASE : List[Any] = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert("RGB" ).resize((256, 256) ) if str(_lowerCAmelCase ).startswith("mps" ): __SCREAMING_SNAKE_CASE : List[str] = torch.manual_seed(_lowerCAmelCase ) else: __SCREAMING_SNAKE_CASE : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : List[str] = { 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 10, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def a_ ( self ): __SCREAMING_SNAKE_CASE : Optional[int] = 'cpu' __SCREAMING_SNAKE_CASE : Dict = self.get_dummy_components() __SCREAMING_SNAKE_CASE : List[str] = self.pipeline_class(**_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Optional[int] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : Dict = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) __SCREAMING_SNAKE_CASE : int = output.images __SCREAMING_SNAKE_CASE : List[Any] = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] __SCREAMING_SNAKE_CASE : List[str] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE : Any = np.array( [0.6199778, 0.63984406, 0.46145785, 0.62944984, 0.5622215, 0.47306132, 0.47441456, 0.4607606, 0.48719263] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def a_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self ): __SCREAMING_SNAKE_CASE : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_img2img_frog.npy" ) __SCREAMING_SNAKE_CASE : List[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) __SCREAMING_SNAKE_CASE : List[Any] = 'A red cartoon frog, 4k' __SCREAMING_SNAKE_CASE : Optional[int] = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : str = KandinskyVaaImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) __SCREAMING_SNAKE_CASE : Dict = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) __SCREAMING_SNAKE_CASE : str = torch.Generator(device="cpu" ).manual_seed(0 ) __SCREAMING_SNAKE_CASE : Optional[int] = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt="" , ).to_tuple() __SCREAMING_SNAKE_CASE : int = pipeline( image=_lowerCAmelCase , image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=100 , height=768 , width=768 , strength=0.2 , output_type="np" , ) __SCREAMING_SNAKE_CASE : str = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
211
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Tuple = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} # See all BART models at https://huggingface.co/models?filter=bart _UpperCAmelCase : Tuple = { '''vocab_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/vocab.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/vocab.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json''', }, '''merges_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/merges.txt''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/merges.txt''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json''', }, } _UpperCAmelCase : Optional[int] = { '''facebook/bart-base''': 10_24, '''facebook/bart-large''': 10_24, '''facebook/bart-large-mnli''': 10_24, '''facebook/bart-large-cnn''': 10_24, '''facebook/bart-large-xsum''': 10_24, '''yjernite/bart_eli5''': 10_24, } class __magic_name__ ( __snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ["input_ids", "attention_mask"] UpperCamelCase__ = BartTokenizer def __init__( self , snake_case_=None , snake_case_=None , snake_case_=None , snake_case_="replace" , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<s>" , snake_case_="<unk>" , snake_case_="<pad>" , snake_case_="<mask>" , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__( _lowerCAmelCase , _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , errors=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , unk_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase , **_lowerCAmelCase , ) lowercase =json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , _lowerCAmelCase ) != add_prefix_space: lowercase =getattr(_lowerCAmelCase , pre_tok_state.pop('''type''' ) ) lowercase =add_prefix_space lowercase =pre_tok_class(**_lowerCAmelCase ) lowercase =add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase ='post_processor' lowercase =getattr(self.backend_tokenizer , _lowerCAmelCase , _lowerCAmelCase ) if tokenizer_component_instance: lowercase =json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase =tuple(state['''sep'''] ) if "cls" in state: lowercase =tuple(state['''cls'''] ) lowercase =False if state.get('''add_prefix_space''' , _lowerCAmelCase ) != add_prefix_space: lowercase =add_prefix_space lowercase =True if state.get('''trim_offsets''' , _lowerCAmelCase ) != trim_offsets: lowercase =trim_offsets lowercase =True if changes_to_apply: lowercase =getattr(_lowerCAmelCase , state.pop('''type''' ) ) lowercase =component_class(**_lowerCAmelCase ) setattr(self.backend_tokenizer , _lowerCAmelCase , _lowerCAmelCase ) @property def _A( self ): if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def _A( self , snake_case_ ): lowercase =AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else value lowercase =value def _A( self , *snake_case_ , **snake_case_ ): lowercase =kwargs.get('''is_split_into_words''' , _lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , *snake_case_ , **snake_case_ ): lowercase =kwargs.get('''is_split_into_words''' , _lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*_lowerCAmelCase , **_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_ = None ): lowercase =self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase ) def _A( self , snake_case_ , snake_case_=None ): lowercase =[self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): lowercase =[self.sep_token_id] lowercase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
72
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
0
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer UpperCamelCase__ = ['bert-base-uncased', 'bert-base-cased'] UpperCamelCase__ = 'hf-internal-testing/tiny-bert-tf-only' if is_tf_available(): class A ( tf.keras.Model ): def __init__(self : Any , __UpperCAmelCase : List[Any] ) -> Optional[Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = tokenizer UpperCAmelCase__ = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCAmelCase__ = TFAutoModel.from_config(_lowerCAmelCase ) def lowercase_ (self : Dict , __UpperCAmelCase : Any ) -> str: """simple docstring""" UpperCAmelCase__ = self.tokenizer(_lowerCAmelCase ) UpperCAmelCase__ = self.bert(**_lowerCAmelCase ) return out["pooler_output"] @require_tf @require_tensorflow_text class A ( unittest.TestCase ): def lowercase_ (self : List[str] ) -> Any: """simple docstring""" super().setUp() UpperCAmelCase__ = [ BertTokenizer.from_pretrained(_lowerCAmelCase ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false UpperCAmelCase__ = [TFBertTokenizer.from_pretrained(_lowerCAmelCase ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(_lowerCAmelCase , use_fast_bert_tokenizer=_lowerCAmelCase ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCAmelCase__ = [ 'This is a straightforward English test sentence.', 'This one has some weird characters\rto\nsee\r\nif those\u00E9break things.', 'Now we\'re going to add some Chinese: 一 二 三 一二三', 'And some much more rare Chinese: 齉 堃 齉堃', 'Je vais aussi écrire en français pour tester les accents', 'Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ', ] UpperCAmelCase__ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def lowercase_ (self : int ) -> Union[str, Any]: """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase__ = tokenizer(_lowerCAmelCase , return_tensors="tf" , padding="longest" ) UpperCAmelCase__ = tf_tokenizer(_lowerCAmelCase ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def lowercase_ (self : Union[str, Any] ) -> List[str]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = tf_tokenizer(self.paired_sentences ) UpperCAmelCase__ = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = tf.function(_lowerCAmelCase ) for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase__ = tf.constant(_lowerCAmelCase ) UpperCAmelCase__ = compiled_tokenizer(_lowerCAmelCase ) UpperCAmelCase__ = tf_tokenizer(_lowerCAmelCase ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def lowercase_ (self : Tuple ) -> int: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = ModelToSave(tokenizer=_lowerCAmelCase ) UpperCAmelCase__ = tf.convert_to_tensor(self.test_sentences ) UpperCAmelCase__ = model(_lowerCAmelCase ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCAmelCase__ = Path(_lowerCAmelCase ) / 'saved.model' model.save(_lowerCAmelCase ) UpperCAmelCase__ = tf.keras.models.load_model(_lowerCAmelCase ) UpperCAmelCase__ = loaded_model(_lowerCAmelCase ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
486
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
0
import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class snake_case__ : """simple docstring""" def __init__( self , __lowercase , __lowercase=3 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=False , __lowercase=True , __lowercase=9_9 , __lowercase=3_2 , __lowercase=5 , __lowercase=4 , __lowercase=3_7 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_1_2 , __lowercase=1_6 , __lowercase=2 , __lowercase=0.0_2 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> Union[str, Any]: """simple docstring""" a__ : Dict = parent a__ : Optional[int] = batch_size a__ : List[str] = seq_length a__ : List[str] = is_training a__ : Optional[int] = use_input_mask a__ : Union[str, Any] = use_token_type_ids a__ : List[str] = use_labels a__ : Tuple = vocab_size a__ : Tuple = hidden_size a__ : Any = num_hidden_layers a__ : Union[str, Any] = num_attention_heads a__ : Dict = intermediate_size a__ : List[str] = hidden_act a__ : Any = hidden_dropout_prob a__ : Dict = attention_probs_dropout_prob a__ : Optional[int] = max_position_embeddings a__ : Optional[Any] = type_vocab_size a__ : str = type_sequence_label_size a__ : str = initializer_range a__ : List[Any] = num_labels a__ : List[str] = num_choices a__ : Union[str, Any] = scope def SCREAMING_SNAKE_CASE__( self ) -> List[Any]: """simple docstring""" a__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : str = None if self.use_input_mask: a__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) a__ : str = None a__ : List[str] = None a__ : Optional[int] = None a__ : List[str] = None if self.use_labels: a__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) a__ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=_lowerCAmelCase , ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[Any]: """simple docstring""" a__ : Dict = FalconModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Optional[Any] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) a__ : str = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Optional[Any]: """simple docstring""" a__ : Union[str, Any] = True a__ : str = FalconModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Any = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , ) a__ : List[Any] = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , ) a__ : Tuple = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Optional[Any]: """simple docstring""" a__ : Union[str, Any] = FalconForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Tuple = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Optional[int]: """simple docstring""" a__ : Optional[Any] = True a__ : Dict = True a__ : Optional[int] = FalconForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() # first forward pass a__ : Tuple = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase , ) a__ : Optional[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids a__ : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size ) a__ : List[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and a__ : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) a__ : Optional[int] = torch.cat([input_mask, next_mask] , dim=-1 ) a__ : Tuple = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , )['hidden_states'][0] a__ : Any = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , )['hidden_states'][0] # select random slice a__ : Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() a__ : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx].detach() a__ : Tuple = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" a__ : Optional[int] = self.prepare_config_and_inputs() ( a__ ) : Union[str, Any] = config_and_inputs a__ : List[Any] = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class snake_case__ (__snake_case , __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __lowerCAmelCase :Union[str, Any] = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __lowerCAmelCase :Optional[Any] = (FalconForCausalLM,) if is_torch_available() else () __lowerCAmelCase :str = ( { "feature-extraction": FalconModel, "text-classification": FalconForSequenceClassification, "text-generation": FalconForCausalLM, "question-answering": FalconForQuestionAnswering, "token-classification": FalconForTokenClassification, "zero-shot": FalconForSequenceClassification, } if is_torch_available() else {} ) __lowerCAmelCase :Tuple = False __lowerCAmelCase :str = False def SCREAMING_SNAKE_CASE__( self ) -> Optional[int]: """simple docstring""" a__ : Dict = FalconModelTester(self ) a__ : Dict = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" a__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" a__ : Dict = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: a__ : Union[str, Any] = alibi self.model_tester.create_and_check_model(_lowerCAmelCase , *_lowerCAmelCase ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() a__ : Optional[Any] = 3 a__ : int = input_dict['input_ids'] a__ : List[Any] = input_ids.ne(1 ).to(_lowerCAmelCase ) a__ : Optional[int] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) a__ : List[str] = FalconForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Tuple = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Any = self.model_tester.prepare_config_and_inputs_for_common() a__ : List[Any] = 3 a__ : Tuple = 'single_label_classification' a__ : Dict = input_dict['input_ids'] a__ : Union[str, Any] = input_ids.ne(1 ).to(_lowerCAmelCase ) a__ : Dict = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) a__ : Tuple = FalconForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Optional[Any] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() a__ : Optional[Any] = input_dict['input_ids'] a__ : Union[str, Any] = FalconForCausalLM(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : List[Any] = model(_lowerCAmelCase , use_cache=_lowerCAmelCase ) a__ : int = input_ids.shape[0] a__ : str = model._convert_to_rw_cache(result.past_key_values ) a__ : List[str] = model._convert_cache_to_standard_format(_lowerCAmelCase , _lowerCAmelCase ) for layer in range(len(_lowerCAmelCase ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() a__ : List[str] = 3 a__ : List[Any] = 'multi_label_classification' a__ : Union[str, Any] = input_dict['input_ids'] a__ : int = input_ids.ne(1 ).to(_lowerCAmelCase ) a__ : List[str] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) a__ : Dict = FalconForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() a__ : Optional[Any] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" for model_class in self.all_generative_model_classes: a__ : Any = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(_lowerCAmelCase , """use_cache""" ): return a__ : Any = model_class(_lowerCAmelCase ).to(_lowerCAmelCase ) if "use_cache" not in inputs: a__ : Optional[int] = True a__ : Tuple = model(**_lowerCAmelCase ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return a__ : str = ( getattr(_lowerCAmelCase , """decoder_layers""" , _lowerCAmelCase ) or getattr(_lowerCAmelCase , """num_decoder_layers""" , _lowerCAmelCase ) or config.num_hidden_layers ) a__ : Tuple = getattr(_lowerCAmelCase , """num_kv_heads""" , config.num_attention_heads ) a__ : Any = getattr(_lowerCAmelCase , """d_model""" , config.hidden_size ) a__ : List[Any] = embed_dim // num_attention_heads a__ : List[str] = outputs['past_key_values'] self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) a__ : Optional[Any] = inputs['input_ids'].shape for i in range(_lowerCAmelCase ): if config.new_decoder_architecture: a__ : Dict = config.num_attention_heads elif config.multi_query: a__ : Dict = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class snake_case__ (unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE__( self ) -> Optional[int]: """simple docstring""" a__ : List[str] = AutoTokenizer.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) a__ : Any = FalconForCausalLM.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) model.eval() model.to(_lowerCAmelCase ) a__ : Union[str, Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(_lowerCAmelCase ) a__ : Tuple = ( 'My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday.' ) a__ : Optional[Any] = model.generate(**_lowerCAmelCase , do_sample=_lowerCAmelCase , max_new_tokens=1_9 ) a__ : Tuple = tokenizer.batch_decode(_lowerCAmelCase )[0] self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: a__ : List[str] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) a__ : str = FalconForCausalLM.from_pretrained(_lowerCAmelCase ) model.eval() model.to(_lowerCAmelCase ) a__ : str = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(_lowerCAmelCase ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**_lowerCAmelCase , do_sample=_lowerCAmelCase , max_new_tokens=4 ) model.generate(**_lowerCAmelCase , do_sample=_lowerCAmelCase , max_new_tokens=4 ) model.generate(**_lowerCAmelCase , num_beams=2 , max_new_tokens=4 ) @slow def SCREAMING_SNAKE_CASE__( self ) -> List[Any]: """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: a__ : str = AutoTokenizer.from_pretrained(_lowerCAmelCase ) a__ : Union[str, Any] = FalconForCausalLM.from_pretrained(_lowerCAmelCase ) model.eval() model.to(device=_lowerCAmelCase ) a__ : Any = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(_lowerCAmelCase ) # Test results are the same with and without cache a__ : str = model.generate(**_lowerCAmelCase , do_sample=_lowerCAmelCase , max_new_tokens=2_0 , use_cache=_lowerCAmelCase ) a__ : Dict = model.generate(**_lowerCAmelCase , do_sample=_lowerCAmelCase , max_new_tokens=2_0 , use_cache=_lowerCAmelCase ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
136
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase : List[Any] = { "configuration_jukebox": [ "JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP", "JukeboxConfig", "JukeboxPriorConfig", "JukeboxVQVAEConfig", ], "tokenization_jukebox": ["JukeboxTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Dict = [ "JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST", "JukeboxModel", "JukeboxPreTrainedModel", "JukeboxVQVAE", "JukeboxPrior", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
327
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
0
'''simple docstring''' def a ( lowerCamelCase__ = 2_00 ): '''simple docstring''' A_ : Optional[Any] = [1, 2, 5, 10, 20, 50, 1_00, 2_00] A_ : List[str] = [0] * (pence + 1) A_ : Any = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowerCamelCase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(2_0_0) == 7_3_6_8_2
667
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
0
import math def SCREAMING_SNAKE_CASE__ ( snake_case__ :int = 100 ) -> int: _lowercase = sum(i * i for i in range(1 , n + 1 ) ) _lowercase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
67
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = '''blenderbot-small''' SCREAMING_SNAKE_CASE_ : int = ['''past_key_values'''] SCREAMING_SNAKE_CASE_ : Tuple = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] ,__A : List[Any]=5_0265 ,__A : str=512 ,__A : Optional[int]=8 ,__A : Any=2048 ,__A : Tuple=16 ,__A : str=8 ,__A : int=2048 ,__A : List[str]=16 ,__A : Optional[int]=0.0 ,__A : Any=0.0 ,__A : int=True ,__A : List[Any]=True ,__A : Tuple="gelu" ,__A : Any=512 ,__A : Dict=0.1 ,__A : Tuple=0.0 ,__A : int=0.0 ,__A : int=0.02 ,__A : Dict=1 ,__A : str=False ,__A : Dict=0 ,__A : Union[str, Any]=1 ,__A : Optional[int]=2 ,__A : List[str]=2 ,**__A : Tuple ,) -> Tuple: _lowercase = vocab_size _lowercase = max_position_embeddings _lowercase = d_model _lowercase = encoder_ffn_dim _lowercase = encoder_layers _lowercase = encoder_attention_heads _lowercase = decoder_ffn_dim _lowercase = decoder_layers _lowercase = decoder_attention_heads _lowercase = dropout _lowercase = attention_dropout _lowercase = activation_dropout _lowercase = activation_function _lowercase = init_std _lowercase = encoder_layerdrop _lowercase = decoder_layerdrop _lowercase = use_cache _lowercase = encoder_layers _lowercase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=__A ,bos_token_id=__A ,eos_token_id=__A ,is_encoder_decoder=__A ,decoder_start_token_id=__A ,forced_eos_token_id=__A ,**__A ,) class A_ ( UpperCAmelCase ): """simple docstring""" @property def __UpperCAmelCase ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _lowercase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: _lowercase = {0: 'batch'} _lowercase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: _lowercase = {0: 'batch', 1: 'decoder_sequence'} _lowercase = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(__A ,direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. _lowercase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: _lowercase , _lowercase = self.num_layers for i in range(__A ): _lowercase = {0: 'batch', 2: 'past_sequence + sequence'} _lowercase = {0: 'batch', 2: 'past_sequence + sequence'} else: _lowercase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def __UpperCAmelCase ( self : Dict ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _lowercase = super().outputs else: _lowercase = super(__A ,self ).outputs if self.use_past: _lowercase , _lowercase = self.num_layers for i in range(__A ): _lowercase = {0: 'batch', 2: 'past_sequence + sequence'} _lowercase = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def __UpperCAmelCase ( self : Optional[int] ,__A : PreTrainedTokenizer ,__A : int = -1 ,__A : int = -1 ,__A : bool = False ,__A : Optional[TensorType] = None ,) -> Mapping[str, Any]: _lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A ,__A ,__A ,__A ,__A ) # Generate decoder inputs _lowercase = seq_length if not self.use_past else 1 _lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A ,__A ,__A ,__A ,__A ) _lowercase = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} _lowercase = dict(**__A ,**__A ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowercase , _lowercase = common_inputs['input_ids'].shape _lowercase = common_inputs['decoder_input_ids'].shape[1] _lowercase , _lowercase = self.num_attention_heads _lowercase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowercase = decoder_seq_length + 3 _lowercase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) _lowercase = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(__A ,__A )] ,dim=1 ) _lowercase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered _lowercase , _lowercase = self.num_layers _lowercase = min(__A ,__A ) _lowercase = max(__A ,__A ) - min_num_layers _lowercase = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(__A ): common_inputs["past_key_values"].append( ( torch.zeros(__A ), torch.zeros(__A ), torch.zeros(__A ), torch.zeros(__A ), ) ) # TODO: test this. _lowercase = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(__A ,__A ): common_inputs["past_key_values"].append((torch.zeros(__A ), torch.zeros(__A )) ) return common_inputs def __UpperCAmelCase ( self : List[Any] ,__A : PreTrainedTokenizer ,__A : int = -1 ,__A : int = -1 ,__A : bool = False ,__A : Optional[TensorType] = None ,) -> Mapping[str, Any]: _lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A ,__A ,__A ,__A ,__A ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowercase , _lowercase = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowercase = seqlen + 2 _lowercase , _lowercase = self.num_layers _lowercase , _lowercase = self.num_attention_heads _lowercase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowercase = common_inputs['attention_mask'].dtype _lowercase = torch.cat( [common_inputs['attention_mask'], torch.ones(__A ,__A ,dtype=__A )] ,dim=1 ) _lowercase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(__A ) ] return common_inputs def __UpperCAmelCase ( self : Any ,__A : PreTrainedTokenizer ,__A : int = -1 ,__A : int = -1 ,__A : bool = False ,__A : Optional[TensorType] = None ,) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _lowercase = compute_effective_axis_dimension( __A ,fixed_dimension=OnnxConfig.default_fixed_batch ,num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _lowercase = tokenizer.num_special_tokens_to_add(__A ) _lowercase = compute_effective_axis_dimension( __A ,fixed_dimension=OnnxConfig.default_fixed_sequence ,num_token_to_add=__A ) # Generate dummy inputs according to compute batch and sequence _lowercase = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size _lowercase = dict(tokenizer(__A ,return_tensors=__A ) ) return common_inputs def __UpperCAmelCase ( self : Dict ,__A : PreTrainedTokenizer ,__A : int = -1 ,__A : int = -1 ,__A : bool = False ,__A : Optional[TensorType] = None ,) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: _lowercase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __A ,batch_size=__A ,seq_length=__A ,is_pair=__A ,framework=__A ) elif self.task == "causal-lm": _lowercase = self._generate_dummy_inputs_for_causal_lm( __A ,batch_size=__A ,seq_length=__A ,is_pair=__A ,framework=__A ) else: _lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A ,batch_size=__A ,seq_length=__A ,is_pair=__A ,framework=__A ) return common_inputs def __UpperCAmelCase ( self : List[str] ,__A : Dict ,__A : Any ,__A : List[Any] ,__A : Tuple ) -> Union[str, Any]: if self.task in ["default", "seq2seq-lm"]: _lowercase = super()._flatten_past_key_values_(__A ,__A ,__A ,__A ) else: _lowercase = super(__A ,self )._flatten_past_key_values_( __A ,__A ,__A ,__A )
67
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
67
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class A_ ( unittest.TestCase ): """simple docstring""" @property def __UpperCAmelCase ( self : Any ) -> str: torch.manual_seed(0 ) _lowercase = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __UpperCAmelCase ( self : Union[str, Any] ) -> Tuple: _lowercase = self.dummy_uncond_unet _lowercase = KarrasVeScheduler() _lowercase = KarrasVePipeline(unet=__A ,scheduler=__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _lowercase = torch.manual_seed(0 ) _lowercase = pipe(num_inference_steps=2 ,generator=__A ,output_type='numpy' ).images _lowercase = torch.manual_seed(0 ) _lowercase = pipe(num_inference_steps=2 ,generator=__A ,output_type='numpy' ,return_dict=__A )[0] _lowercase = image[0, -3:, -3:, -1] _lowercase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowercase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class A_ ( unittest.TestCase ): """simple docstring""" def __UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: _lowercase = 'google/ncsnpp-celebahq-256' _lowercase = UNetaDModel.from_pretrained(__A ) _lowercase = KarrasVeScheduler() _lowercase = KarrasVePipeline(unet=__A ,scheduler=__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _lowercase = torch.manual_seed(0 ) _lowercase = pipe(num_inference_steps=20 ,generator=__A ,output_type='numpy' ).images _lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _lowercase = np.array([0.578, 0.5811, 0.5924, 0.5809, 0.587, 0.5886, 0.5861, 0.5802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
67
1
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : torch.FloatTensor SCREAMING_SNAKE_CASE_ : Optional[torch.FloatTensor] = None def SCREAMING_SNAKE_CASE__ ( snake_case__ :Tuple , snake_case__ :List[str]=0.999 , snake_case__ :Dict="cosine" , ) -> List[Any]: if alpha_transform_type == "cosine": def alpha_bar_fn(snake_case__ :Optional[int] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(snake_case__ :Optional[int] ): return math.exp(t * -12.0 ) else: raise ValueError(F"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) _lowercase = [] for i in range(snake_case__ ): _lowercase = i / num_diffusion_timesteps _lowercase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(snake_case__ ) / alpha_bar_fn(snake_case__ ) , snake_case__ ) ) return torch.tensor(snake_case__ , dtype=torch.floataa ) class A_ ( UpperCAmelCase , UpperCAmelCase ): """simple docstring""" @register_to_config def __init__( self : Union[str, Any] ,__A : int = 1000 ,__A : str = "fixed_small_log" ,__A : bool = True ,__A : Optional[float] = 1.0 ,__A : str = "epsilon" ,__A : str = "squaredcos_cap_v2" ,) -> int: if beta_schedule != "squaredcos_cap_v2": raise ValueError('UnCLIPScheduler only supports `beta_schedule`: \'squaredcos_cap_v2\'' ) _lowercase = betas_for_alpha_bar(__A ) _lowercase = 1.0 - self.betas _lowercase = torch.cumprod(self.alphas ,dim=0 ) _lowercase = torch.tensor(1.0 ) # standard deviation of the initial noise distribution _lowercase = 1.0 # setable values _lowercase = None _lowercase = torch.from_numpy(np.arange(0 ,__A )[::-1].copy() ) _lowercase = variance_type def __UpperCAmelCase ( self : List[str] ,__A : torch.FloatTensor ,__A : Optional[int] = None ) -> torch.FloatTensor: return sample def __UpperCAmelCase ( self : List[str] ,__A : int ,__A : Union[str, torch.device] = None ) -> List[str]: _lowercase = num_inference_steps _lowercase = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) _lowercase = (np.arange(0 ,__A ) * step_ratio).round()[::-1].copy().astype(np.intaa ) _lowercase = torch.from_numpy(__A ).to(__A ) def __UpperCAmelCase ( self : Dict ,__A : List[Any] ,__A : str=None ,__A : str=None ,__A : List[str]=None ) -> List[str]: if prev_timestep is None: _lowercase = t - 1 _lowercase = self.alphas_cumprod[t] _lowercase = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one _lowercase = 1 - alpha_prod_t _lowercase = 1 - alpha_prod_t_prev if prev_timestep == t - 1: _lowercase = self.betas[t] else: _lowercase = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _lowercase = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: _lowercase = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": _lowercase = torch.log(torch.clamp(__A ,min=1e-20 ) ) _lowercase = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler _lowercase = variance.log() _lowercase = beta.log() _lowercase = (predicted_variance + 1) / 2 _lowercase = frac * max_log + (1 - frac) * min_log return variance def __UpperCAmelCase ( self : Any ,__A : torch.FloatTensor ,__A : int ,__A : torch.FloatTensor ,__A : Optional[int] = None ,__A : Optional[int]=None ,__A : bool = True ,) -> Union[UnCLIPSchedulerOutput, Tuple]: _lowercase = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": _lowercase , _lowercase = torch.split(__A ,sample.shape[1] ,dim=1 ) else: _lowercase = None # 1. compute alphas, betas if prev_timestep is None: _lowercase = t - 1 _lowercase = self.alphas_cumprod[t] _lowercase = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one _lowercase = 1 - alpha_prod_t _lowercase = 1 - alpha_prod_t_prev if prev_timestep == t - 1: _lowercase = self.betas[t] _lowercase = self.alphas[t] else: _lowercase = 1 - alpha_prod_t / alpha_prod_t_prev _lowercase = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _lowercase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _lowercase = model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" ' for the UnCLIPScheduler.' ) # 3. Clip "predicted x_0" if self.config.clip_sample: _lowercase = torch.clamp( __A ,-self.config.clip_sample_range ,self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _lowercase = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t _lowercase = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _lowercase = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _lowercase = 0 if t > 0: _lowercase = randn_tensor( model_output.shape ,dtype=model_output.dtype ,generator=__A ,device=model_output.device ) _lowercase = self._get_variance( __A ,predicted_variance=__A ,prev_timestep=__A ,) if self.variance_type == "fixed_small_log": _lowercase = variance elif self.variance_type == "learned_range": _lowercase = (0.5 * variance).exp() else: raise ValueError( F"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" ' for the UnCLIPScheduler.' ) _lowercase = variance * variance_noise _lowercase = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__A ,pred_original_sample=__A ) def __UpperCAmelCase ( self : Tuple ,__A : torch.FloatTensor ,__A : torch.FloatTensor ,__A : torch.IntTensor ,) -> torch.FloatTensor: # Make sure alphas_cumprod and timestep have same device and dtype as original_samples _lowercase = self.alphas_cumprod.to(device=original_samples.device ,dtype=original_samples.dtype ) _lowercase = timesteps.to(original_samples.device ) _lowercase = alphas_cumprod[timesteps] ** 0.5 _lowercase = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): _lowercase = sqrt_alpha_prod.unsqueeze(-1 ) _lowercase = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): _lowercase = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) _lowercase = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
67
def SCREAMING_SNAKE_CASE__ ( snake_case__ :str , snake_case__ :str ) -> list: _lowercase = len(snake_case__ ) _lowercase = [] for i in range(len(snake_case__ ) - pat_len + 1 ): _lowercase = True for j in range(snake_case__ ): if s[i + j] != pattern[j]: _lowercase = False break if match_found: position.append(snake_case__ ) return position if __name__ == "__main__": assert naive_pattern_search("""ABCDEFG""", """DE""") == [3] print(naive_pattern_search("""ABAAABCDBBABCDDEBCABC""", """ABC"""))
67
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import FEATURE_EXTRACTOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor, ChineseCLIPProcessor @require_vision class A_ ( unittest.TestCase ): """simple docstring""" def __UpperCAmelCase ( self : Optional[int] ) -> List[str]: _lowercase = tempfile.mkdtemp() _lowercase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', '的', '价', '格', '是', '15', '便', 'alex', '##andra', ',', '。', '-', 't', 'shirt', ] _lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file ,'w' ,encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) _lowercase = { 'do_resize': True, 'size': {'height': 224, 'width': 224}, 'do_center_crop': True, 'crop_size': {'height': 18, 'width': 18}, 'do_normalize': True, 'image_mean': [0.48145466, 0.4578275, 0.40821073], 'image_std': [0.26862954, 0.26130258, 0.27577711], 'do_convert_rgb': True, } _lowercase = os.path.join(self.tmpdirname ,__A ) with open(self.image_processor_file ,'w' ,encoding='utf-8' ) as fp: json.dump(__A ,__A ) def __UpperCAmelCase ( self : Optional[int] ,**__A : Tuple ) -> Any: return BertTokenizer.from_pretrained(self.tmpdirname ,**__A ) def __UpperCAmelCase ( self : int ,**__A : Union[str, Any] ) -> List[Any]: return BertTokenizerFast.from_pretrained(self.tmpdirname ,**__A ) def __UpperCAmelCase ( self : List[Any] ,**__A : List[Any] ) -> List[str]: return ChineseCLIPImageProcessor.from_pretrained(self.tmpdirname ,**__A ) def __UpperCAmelCase ( self : Tuple ) -> Dict: shutil.rmtree(self.tmpdirname ) def __UpperCAmelCase ( self : Tuple ) -> int: _lowercase = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] _lowercase = [Image.fromarray(np.moveaxis(__A ,0 ,-1 ) ) for x in image_inputs] return image_inputs def __UpperCAmelCase ( self : Optional[int] ) -> List[str]: _lowercase = self.get_tokenizer() _lowercase = self.get_rust_tokenizer() _lowercase = self.get_image_processor() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) processor_slow.save_pretrained(self.tmpdirname ) _lowercase = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ,use_fast=__A ) _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) processor_fast.save_pretrained(self.tmpdirname ) _lowercase = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() ,tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() ,tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() ,tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer ,__A ) self.assertIsInstance(processor_fast.tokenizer ,__A ) self.assertEqual(processor_slow.image_processor.to_json_string() ,image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() ,image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor ,__A ) self.assertIsInstance(processor_fast.image_processor ,__A ) def __UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: _lowercase = ChineseCLIPProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _lowercase = self.get_tokenizer(cls_token='(CLS)' ,sep_token='(SEP)' ) _lowercase = self.get_image_processor(do_normalize=__A ) _lowercase = ChineseCLIPProcessor.from_pretrained( self.tmpdirname ,cls_token='(CLS)' ,sep_token='(SEP)' ,do_normalize=__A ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,__A ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,__A ) def __UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: _lowercase = self.get_image_processor() _lowercase = self.get_tokenizer() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) _lowercase = self.prepare_image_inputs() _lowercase = image_processor(__A ,return_tensors='np' ) _lowercase = processor(images=__A ,return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __UpperCAmelCase ( self : Dict ) -> Union[str, Any]: _lowercase = self.get_image_processor() _lowercase = self.get_tokenizer() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) _lowercase = 'Alexandra,T-shirt的价格是15便士。' _lowercase = processor(text=__A ) _lowercase = tokenizer(__A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __UpperCAmelCase ( self : Optional[int] ) -> str: _lowercase = self.get_image_processor() _lowercase = self.get_tokenizer() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) _lowercase = 'Alexandra,T-shirt的价格是15便士。' _lowercase = self.prepare_image_inputs() _lowercase = processor(text=__A ,images=__A ) self.assertListEqual(list(inputs.keys() ) ,['input_ids', 'token_type_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(__A ): processor() def __UpperCAmelCase ( self : Optional[Any] ) -> int: _lowercase = self.get_image_processor() _lowercase = self.get_tokenizer() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) _lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _lowercase = processor.batch_decode(__A ) _lowercase = tokenizer.batch_decode(__A ) self.assertListEqual(__A ,__A ) def __UpperCAmelCase ( self : Dict ) -> str: _lowercase = self.get_image_processor() _lowercase = self.get_tokenizer() _lowercase = ChineseCLIPProcessor(tokenizer=__A ,image_processor=__A ) _lowercase = 'Alexandra,T-shirt的价格是15便士。' _lowercase = self.prepare_image_inputs() _lowercase = processor(text=__A ,images=__A ) self.assertListEqual(list(inputs.keys() ) ,processor.model_input_names )
67
from typing import Any import numpy as np def SCREAMING_SNAKE_CASE__ ( snake_case__ :np.ndarray ) -> bool: return np.array_equal(snake_case__ , matrix.conjugate().T ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :np.ndarray , snake_case__ :np.ndarray ) -> Any: _lowercase = v.conjugate().T _lowercase = v_star.dot(snake_case__ ) assert isinstance(snake_case__ , np.ndarray ) return (v_star_dot.dot(snake_case__ )) / (v_star.dot(snake_case__ )) def SCREAMING_SNAKE_CASE__ ( ) -> None: _lowercase = np.array([[2, 2 + 1J, 4], [2 - 1J, 3, 1J], [4, -1J, 1]] ) _lowercase = np.array([[1], [2], [3]] ) assert is_hermitian(snake_case__ ), F"""{a} is not hermitian.""" print(rayleigh_quotient(snake_case__ , snake_case__ ) ) _lowercase = np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(snake_case__ ), F"""{a} is not hermitian.""" assert rayleigh_quotient(snake_case__ , snake_case__ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
67
1
from collections import deque from .hash_table import HashTable class A_ ( UpperCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] ,*__A : Optional[int] ,**__A : List[str] ) -> List[str]: super().__init__(*__A ,**__A ) def __UpperCAmelCase ( self : Any ,__A : List[Any] ,__A : List[str] ) -> int: _lowercase = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(__A ) _lowercase = self.values[key] def __UpperCAmelCase ( self : Any ) -> List[Any]: return ( sum(self.charge_factor - len(__A ) for slot in self.values ) / self.size_table * self.charge_factor ) def __UpperCAmelCase ( self : Dict ,__A : Optional[Any] ,__A : str=None ) -> List[Any]: if not ( len(self.values[key] ) == self.charge_factor and self.values.count(__A ) == 0 ): return key return super()._collision_resolution(__A ,__A )
67
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class A_ ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple ,__A : Dict ,__A : List[Any]=7 ,__A : Dict=3 ,__A : Tuple=30 ,__A : Dict=400 ,__A : Any=True ,__A : List[Any]=None ,__A : Any=True ,__A : List[str]=[0.5, 0.5, 0.5] ,__A : Union[str, Any]=[0.5, 0.5, 0.5] ,__A : int=True ,__A : List[str]=1 / 255 ,__A : Union[str, Any]=True ,) -> List[Any]: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowercase = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} _lowercase = parent _lowercase = batch_size _lowercase = num_channels _lowercase = min_resolution _lowercase = max_resolution _lowercase = do_resize _lowercase = size _lowercase = do_normalize _lowercase = image_mean _lowercase = image_std _lowercase = do_rescale _lowercase = rescale_factor _lowercase = do_pad def __UpperCAmelCase ( self : str ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __UpperCAmelCase ( self : Tuple ,__A : Union[str, Any] ,__A : List[str]=False ) -> Union[str, Any]: if not batched: _lowercase = image_inputs[0] if isinstance(__A ,Image.Image ): _lowercase , _lowercase = image.size else: _lowercase , _lowercase = image.shape[1], image.shape[2] if w < h: _lowercase = int(self.size['shortest_edge'] * h / w ) _lowercase = self.size['shortest_edge'] elif w > h: _lowercase = self.size['shortest_edge'] _lowercase = int(self.size['shortest_edge'] * w / h ) else: _lowercase = self.size['shortest_edge'] _lowercase = self.size['shortest_edge'] else: _lowercase = [] for image in image_inputs: _lowercase , _lowercase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowercase = max(__A ,key=lambda __A : item[0] )[0] _lowercase = max(__A ,key=lambda __A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A_ ( UpperCAmelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = DetaImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self : List[Any] ) -> Tuple: _lowercase = DetaImageProcessingTester(self ) @property def __UpperCAmelCase ( self : List[Any] ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self : Tuple ) -> List[Any]: _lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A ,'image_mean' ) ) self.assertTrue(hasattr(__A ,'image_std' ) ) self.assertTrue(hasattr(__A ,'do_normalize' ) ) self.assertTrue(hasattr(__A ,'do_resize' ) ) self.assertTrue(hasattr(__A ,'do_rescale' ) ) self.assertTrue(hasattr(__A ,'do_pad' ) ) self.assertTrue(hasattr(__A ,'size' ) ) def __UpperCAmelCase ( self : str ) -> List[str]: _lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad ,__A ) def __UpperCAmelCase ( self : List[Any] ) -> Any: pass def __UpperCAmelCase ( self : Optional[int] ) -> Tuple: # Initialize image_processing _lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A ,Image.Image ) # Test not batched input _lowercase = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ,batched=__A ) _lowercase = image_processing(__A ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def __UpperCAmelCase ( self : Union[str, Any] ) -> List[str]: # Initialize image_processing _lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ,numpify=__A ) for image in image_inputs: self.assertIsInstance(__A ,np.ndarray ) # Test not batched input _lowercase = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched _lowercase = image_processing(__A ,return_tensors='pt' ).pixel_values _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ,batched=__A ) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def __UpperCAmelCase ( self : Optional[Any] ) -> Any: # Initialize image_processing _lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ,torchify=__A ) for image in image_inputs: self.assertIsInstance(__A ,torch.Tensor ) # Test not batched input _lowercase = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched _lowercase = image_processing(__A ,return_tensors='pt' ).pixel_values _lowercase , _lowercase = self.image_processor_tester.get_expected_values(__A ,batched=__A ) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) @slow def __UpperCAmelCase ( self : List[str] ) -> List[str]: # prepare image and target _lowercase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' ,'r' ) as f: _lowercase = json.loads(f.read() ) _lowercase = {'image_id': 3_9769, 'annotations': target} # encode them _lowercase = DetaImageProcessor() _lowercase = image_processing(images=__A ,annotations=__A ,return_tensors='pt' ) # verify pixel values _lowercase = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape ,__A ) _lowercase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] ,__A ,atol=1e-4 ) ) # verify area _lowercase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] ,__A ) ) # verify boxes _lowercase = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape ,__A ) _lowercase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] ,__A ,atol=1e-3 ) ) # verify image_id _lowercase = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] ,__A ) ) # verify is_crowd _lowercase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] ,__A ) ) # verify class_labels _lowercase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] ,__A ) ) # verify orig_size _lowercase = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] ,__A ) ) # verify size _lowercase = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] ,__A ) ) @slow def __UpperCAmelCase ( self : List[Any] ) -> Dict: # prepare image, target and masks_path _lowercase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' ,'r' ) as f: _lowercase = json.loads(f.read() ) _lowercase = {'file_name': '000000039769.png', 'image_id': 3_9769, 'segments_info': target} _lowercase = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them _lowercase = DetaImageProcessor(format='coco_panoptic' ) _lowercase = image_processing(images=__A ,annotations=__A ,masks_path=__A ,return_tensors='pt' ) # verify pixel values _lowercase = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape ,__A ) _lowercase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] ,__A ,atol=1e-4 ) ) # verify area _lowercase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] ,__A ) ) # verify boxes _lowercase = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape ,__A ) _lowercase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] ,__A ,atol=1e-3 ) ) # verify image_id _lowercase = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] ,__A ) ) # verify is_crowd _lowercase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] ,__A ) ) # verify class_labels _lowercase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] ,__A ) ) # verify masks _lowercase = 82_2873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() ,__A ) # verify orig_size _lowercase = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] ,__A ) ) # verify size _lowercase = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] ,__A ) )
67
1
import heapq import sys import numpy as np snake_case = tuple[int, int] class A_ : """simple docstring""" def __init__( self : List[Any] ) -> List[Any]: _lowercase = [] _lowercase = set() def __UpperCAmelCase ( self : List[str] ) -> Optional[int]: if not self.empty(): return self.elements[0][0] else: return float('inf' ) def __UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: return len(self.elements ) == 0 def __UpperCAmelCase ( self : str ,__A : Any ,__A : Any ) -> int: if item not in self.set: heapq.heappush(self.elements ,(priority, item) ) self.set.add(__A ) else: # update # print("update", item) _lowercase = [] ((_lowercase) , (_lowercase)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((_lowercase) , (_lowercase)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements ,(pro, xxx) ) def __UpperCAmelCase ( self : List[str] ,__A : Optional[Any] ) -> Dict: if item in self.set: self.set.remove(__A ) _lowercase = [] ((_lowercase) , (_lowercase)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((_lowercase) , (_lowercase)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements ,(prito, yyy) ) def __UpperCAmelCase ( self : List[str] ) -> Any: return self.elements[0][1] def __UpperCAmelCase ( self : Any ) -> Any: ((_lowercase) , (_lowercase)) = heapq.heappop(self.elements ) self.set.remove(__A ) return (priority, item) def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos , snake_case__ :TPos ) -> Tuple: # euclidean distance _lowercase = np.array(snake_case__ ) _lowercase = np.array(snake_case__ ) return np.linalg.norm(a - b ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos , snake_case__ :TPos ) -> str: # integer division by time variable return consistent_heuristic(snake_case__ , snake_case__ ) // t def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos , snake_case__ :TPos ) -> Any: # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos , snake_case__ :int , snake_case__ :TPos , snake_case__ :dict[TPos, float] ) -> str: _lowercase = g_function[start] + Wa * heuristics[i](snake_case__ , snake_case__ ) return ans def SCREAMING_SNAKE_CASE__ ( snake_case__ :Union[str, Any] , snake_case__ :List[str] , snake_case__ :Optional[int] ) -> int: _lowercase = np.chararray((n, n) ) for i in range(snake_case__ ): for j in range(snake_case__ ): _lowercase = '*' for i in range(snake_case__ ): for j in range(snake_case__ ): if (j, (n - 1) - i) in blocks: _lowercase = '#' _lowercase = '-' _lowercase = back_pointer[goal] while x != start: ((_lowercase) , (_lowercase)) = x # print(x) _lowercase = '-' _lowercase = back_pointer[x] _lowercase = '-' for i in range(snake_case__ ): for j in range(snake_case__ ): if (i, j) == (0, n - 1): print(grid[i][j] , end=' ' ) print('<-- End position' , end=' ' ) else: print(grid[i][j] , end=' ' ) print() print('^' ) print('Start position' ) print() print('# is an obstacle' ) print('- is the path taken by algorithm' ) print('PATH TAKEN BY THE ALGORITHM IS:-' ) _lowercase = back_pointer[goal] while x != start: print(snake_case__ , end=' ' ) _lowercase = back_pointer[x] print(snake_case__ ) sys.exit() def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos ) -> str: if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def SCREAMING_SNAKE_CASE__ ( snake_case__ :str , snake_case__ :Tuple , snake_case__ :Optional[int] , snake_case__ :str , snake_case__ :Optional[int] , snake_case__ :int , snake_case__ :Tuple , snake_case__ :str , ) -> Any: for itera in range(snake_case__ ): open_list[itera].remove_element(snake_case__ ) # print("s", s) # print("j", j) ((_lowercase) , (_lowercase)) = s _lowercase = (x - 1, y) _lowercase = (x + 1, y) _lowercase = (x, y + 1) _lowercase = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(snake_case__ ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(snake_case__ ) _lowercase = -1 _lowercase = float('inf' ) if valid(snake_case__ ) and g_function[neighbours] > g_function[s] + 1: _lowercase = g_function[s] + 1 _lowercase = s if neighbours not in close_list_anchor: open_list[0].put(snake_case__ , key(snake_case__ , 0 , snake_case__ , snake_case__ ) ) if neighbours not in close_list_inad: for var in range(1 , snake_case__ ): if key(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) <= Wa * key( snake_case__ , 0 , snake_case__ , snake_case__ ): open_list[j].put( snake_case__ , key(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) ) def SCREAMING_SNAKE_CASE__ ( ) -> str: _lowercase = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(15 , 20 ): some_list.append((x, 17) ) for x in range(10 , 19 ): for y in range(1 , 15 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(12 , 19 ): some_list.append((x, y) ) for x in range(3 , 13 ): for y in range(16 , 19 ): some_list.append((x, y) ) return some_list snake_case = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} snake_case = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (1_0, 1), (1_1, 1), (1_2, 1), (1_3, 1), (1_4, 1), (1_5, 1), (1_6, 1), (1_7, 1), (1_8, 1), (1_9, 1), ] snake_case = make_common_ground() snake_case = blocks_blk # hyper parameters snake_case = 1 snake_case = 1 snake_case = 2_0 snake_case = 3 # one consistent and two other inconsistent # start and end destination snake_case = (0, 0) snake_case = (n - 1, n - 1) snake_case = 1 def SCREAMING_SNAKE_CASE__ ( snake_case__ :TPos , snake_case__ :TPos , snake_case__ :int ) -> Any: _lowercase = {start: 0, goal: float('inf' )} _lowercase = {start: -1, goal: -1} _lowercase = [] _lowercase = set() for i in range(snake_case__ ): open_list.append(PriorityQueue() ) open_list[i].put(snake_case__ , key(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) ) _lowercase = [] _lowercase = [] while open_list[0].minkey() < float('inf' ): for i in range(1 , snake_case__ ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float('inf' ): do_something(snake_case__ , snake_case__ , snake_case__ ) else: _lowercase , _lowercase = open_list[i].top_show() visited.add(snake_case__ ) expand_state( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) close_list_inad.append(snake_case__ ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float('inf' ): do_something(snake_case__ , snake_case__ , snake_case__ ) else: _lowercase = open_list[0].top_show() visited.add(snake_case__ ) expand_state( snake_case__ , 0 , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) close_list_anchor.append(snake_case__ ) print('No path found to goal' ) print() for i in range(n - 1 , -1 , -1 ): for j in range(snake_case__ ): if (j, i) in blocks: print('#' , end=' ' ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print('*' , end=' ' ) else: print('-' , end=' ' ) else: print('*' , end=' ' ) if (j, i) == (n - 1, n - 1): print('<-- End position' , end=' ' ) print() print('^' ) print('Start position' ) print() print('# is an obstacle' ) print('- is the path taken by algorithm' ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError("""At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training""") # TF training parameters snake_case = False snake_case = False def SCREAMING_SNAKE_CASE__ ( snake_case__ :Namespace ) -> Tuple: return TrainCommand(snake_case__ ) class A_ ( UpperCAmelCase ): """simple docstring""" @staticmethod def __UpperCAmelCase ( __A : ArgumentParser ) -> List[Any]: _lowercase = parser.add_parser('train' ,help='CLI tool to train a model on a task.' ) train_parser.add_argument( '--train_data' ,type=__A ,required=__A ,help='path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.' ,) train_parser.add_argument( '--column_label' ,type=__A ,default=0 ,help='Column of the dataset csv file with example labels.' ) train_parser.add_argument( '--column_text' ,type=__A ,default=1 ,help='Column of the dataset csv file with example texts.' ) train_parser.add_argument( '--column_id' ,type=__A ,default=2 ,help='Column of the dataset csv file with example ids.' ) train_parser.add_argument( '--skip_first_row' ,action='store_true' ,help='Skip the first row of the csv file (headers).' ) train_parser.add_argument('--validation_data' ,type=__A ,default='' ,help='path to validation dataset.' ) train_parser.add_argument( '--validation_split' ,type=__A ,default=0.1 ,help='if validation dataset is not provided, fraction of train dataset to use as validation dataset.' ,) train_parser.add_argument('--output' ,type=__A ,default='./' ,help='path to saved the trained model.' ) train_parser.add_argument( '--task' ,type=__A ,default='text_classification' ,help='Task to train the model on.' ) train_parser.add_argument( '--model' ,type=__A ,default='bert-base-uncased' ,help='Model\'s name or path to stored model.' ) train_parser.add_argument('--train_batch_size' ,type=__A ,default=32 ,help='Batch size for training.' ) train_parser.add_argument('--valid_batch_size' ,type=__A ,default=64 ,help='Batch size for validation.' ) train_parser.add_argument('--learning_rate' ,type=__A ,default=3e-5 ,help='Learning rate.' ) train_parser.add_argument('--adam_epsilon' ,type=__A ,default=1e-08 ,help='Epsilon for Adam optimizer.' ) train_parser.set_defaults(func=__A ) def __init__( self : Optional[Any] ,__A : Namespace ) -> Tuple: _lowercase = logging.get_logger('transformers-cli/training' ) _lowercase = 'tf' if is_tf_available() else 'torch' os.makedirs(args.output ,exist_ok=__A ) _lowercase = args.output _lowercase = args.column_label _lowercase = args.column_text _lowercase = args.column_id self.logger.info(F"""Loading {args.task} pipeline for {args.model}""" ) if args.task == "text_classification": _lowercase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(F"""Loading dataset from {args.train_data}""" ) _lowercase = Processor.create_from_csv( args.train_data ,column_label=args.column_label ,column_text=args.column_text ,column_id=args.column_id ,skip_first_row=args.skip_first_row ,) _lowercase = None if args.validation_data: self.logger.info(F"""Loading validation dataset from {args.validation_data}""" ) _lowercase = Processor.create_from_csv( args.validation_data ,column_label=args.column_label ,column_text=args.column_text ,column_id=args.column_id ,skip_first_row=args.skip_first_row ,) _lowercase = args.validation_split _lowercase = args.train_batch_size _lowercase = args.valid_batch_size _lowercase = args.learning_rate _lowercase = args.adam_epsilon def __UpperCAmelCase ( self : Optional[Any] ) -> str: if self.framework == "tf": return self.run_tf() return self.run_torch() def __UpperCAmelCase ( self : Tuple ) -> List[Any]: raise NotImplementedError def __UpperCAmelCase ( self : Optional[int] ) -> Optional[int]: self.pipeline.fit( self.train_dataset ,validation_data=self.valid_dataset ,validation_split=self.validation_split ,learning_rate=self.learning_rate ,adam_epsilon=self.adam_epsilon ,train_batch_size=self.train_batch_size ,valid_batch_size=self.valid_batch_size ,) # Save trained pipeline self.pipeline.save_pretrained(self.output )
67
1
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch snake_case = """sshleifer/bart-tiny-random""" snake_case = """patrickvonplaten/t5-tiny-random""" @require_torch class A_ ( unittest.TestCase ): """simple docstring""" @cached_property def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: return AutoConfig.from_pretrained(__A ) def __UpperCAmelCase ( self : List[Any] ) -> Tuple: _lowercase , *_lowercase = create_student_by_copying_alternating_layers(__A ,tempfile.mkdtemp() ,e=1 ,d=1 ) self.assertEqual(student.config.num_hidden_layers ,1 ) def __UpperCAmelCase ( self : Tuple ) -> int: _lowercase , *_lowercase = create_student_by_copying_alternating_layers(__A ,tempfile.mkdtemp() ,e=1 ,d=__A ) def __UpperCAmelCase ( self : Dict ) -> List[Any]: _lowercase , *_lowercase = create_student_by_copying_alternating_layers(__A ,tempfile.mkdtemp() ,e=1 ,d=__A ) self.assertEqual(student.config.encoder_layers ,1 ) self.assertEqual(student.config.decoder_layers ,self.teacher_config.encoder_layers ) def __UpperCAmelCase ( self : List[Any] ) -> Any: _lowercase , *_lowercase = create_student_by_copying_alternating_layers(__A ,tempfile.mkdtemp() ,e=1 ,d=1 ) self.assertEqual(student.config.encoder_layers ,1 ) self.assertEqual(student.config.decoder_layers ,1 ) def __UpperCAmelCase ( self : Optional[int] ) -> Dict: with self.assertRaises(__A ): create_student_by_copying_alternating_layers(__A ,tempfile.mkdtemp() ,e=__A ,d=__A )
67
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( snake_case__ :Any ) -> str: _lowercase = DPTConfig(embedding_type='hybrid' ) if "large" in checkpoint_url: _lowercase = 1024 _lowercase = 4096 _lowercase = 24 _lowercase = 16 _lowercase = [5, 11, 17, 23] _lowercase = [256, 512, 1024, 1024] _lowercase = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: _lowercase = 768 _lowercase = [1, 1, 1, 0.5] _lowercase = [256, 512, 768, 768] _lowercase = 150 _lowercase = 16 _lowercase = (1, 384, 384) _lowercase = False _lowercase = 'project' if "ade" in checkpoint_url: _lowercase = True _lowercase = 768 _lowercase = [1, 1, 1, 0.5] _lowercase = 150 _lowercase = 16 _lowercase = 'huggingface/label-files' _lowercase = 'ade20k-id2label.json' _lowercase = json.load(open(cached_download(hf_hub_url(snake_case__ , snake_case__ , repo_type='dataset' ) ) , 'r' ) ) _lowercase = {int(snake_case__ ): v for k, v in idalabel.items()} _lowercase = idalabel _lowercase = {v: k for k, v in idalabel.items()} _lowercase = [1, 150, 480, 480] return config, expected_shape def SCREAMING_SNAKE_CASE__ ( snake_case__ :str ) -> str: _lowercase = ['pretrained.model.head.weight', 'pretrained.model.head.bias'] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :List[Any] ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): _lowercase = name.replace('pretrained.model' , 'dpt.encoder' ) if "pretrained.model" in name: _lowercase = name.replace('pretrained.model' , 'dpt.embeddings' ) if "patch_embed" in name: _lowercase = name.replace('patch_embed' , '' ) if "pos_embed" in name: _lowercase = name.replace('pos_embed' , 'position_embeddings' ) if "attn.proj" in name: _lowercase = name.replace('attn.proj' , 'attention.output.dense' ) if "proj" in name and "project" not in name: _lowercase = name.replace('proj' , 'projection' ) if "blocks" in name: _lowercase = name.replace('blocks' , 'layer' ) if "mlp.fc1" in name: _lowercase = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _lowercase = name.replace('mlp.fc2' , 'output.dense' ) if "norm1" in name and "backbone" not in name: _lowercase = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name and "backbone" not in name: _lowercase = name.replace('norm2' , 'layernorm_after' ) if "scratch.output_conv" in name: _lowercase = name.replace('scratch.output_conv' , 'head' ) if "scratch" in name: _lowercase = name.replace('scratch' , 'neck' ) if "layer1_rn" in name: _lowercase = name.replace('layer1_rn' , 'convs.0' ) if "layer2_rn" in name: _lowercase = name.replace('layer2_rn' , 'convs.1' ) if "layer3_rn" in name: _lowercase = name.replace('layer3_rn' , 'convs.2' ) if "layer4_rn" in name: _lowercase = name.replace('layer4_rn' , 'convs.3' ) if "refinenet" in name: _lowercase = int(name[len('neck.refinenet' ) : len('neck.refinenet' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 _lowercase = name.replace(F"""refinenet{layer_idx}""" , F"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: _lowercase = name.replace('out_conv' , 'projection' ) if "resConfUnit1" in name: _lowercase = name.replace('resConfUnit1' , 'residual_layer1' ) if "resConfUnit2" in name: _lowercase = name.replace('resConfUnit2' , 'residual_layer2' ) if "conv1" in name: _lowercase = name.replace('conv1' , 'convolution1' ) if "conv2" in name: _lowercase = name.replace('conv2' , 'convolution2' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: _lowercase = name.replace('pretrained.act_postprocess1.0.project.0' , 'neck.reassemble_stage.readout_projects.0.0' ) if "pretrained.act_postprocess2.0.project.0" in name: _lowercase = name.replace('pretrained.act_postprocess2.0.project.0' , 'neck.reassemble_stage.readout_projects.1.0' ) if "pretrained.act_postprocess3.0.project.0" in name: _lowercase = name.replace('pretrained.act_postprocess3.0.project.0' , 'neck.reassemble_stage.readout_projects.2.0' ) if "pretrained.act_postprocess4.0.project.0" in name: _lowercase = name.replace('pretrained.act_postprocess4.0.project.0' , 'neck.reassemble_stage.readout_projects.3.0' ) # resize blocks if "pretrained.act_postprocess1.3" in name: _lowercase = name.replace('pretrained.act_postprocess1.3' , 'neck.reassemble_stage.layers.0.projection' ) if "pretrained.act_postprocess1.4" in name: _lowercase = name.replace('pretrained.act_postprocess1.4' , 'neck.reassemble_stage.layers.0.resize' ) if "pretrained.act_postprocess2.3" in name: _lowercase = name.replace('pretrained.act_postprocess2.3' , 'neck.reassemble_stage.layers.1.projection' ) if "pretrained.act_postprocess2.4" in name: _lowercase = name.replace('pretrained.act_postprocess2.4' , 'neck.reassemble_stage.layers.1.resize' ) if "pretrained.act_postprocess3.3" in name: _lowercase = name.replace('pretrained.act_postprocess3.3' , 'neck.reassemble_stage.layers.2.projection' ) if "pretrained.act_postprocess4.3" in name: _lowercase = name.replace('pretrained.act_postprocess4.3' , 'neck.reassemble_stage.layers.3.projection' ) if "pretrained.act_postprocess4.4" in name: _lowercase = name.replace('pretrained.act_postprocess4.4' , 'neck.reassemble_stage.layers.3.resize' ) if "pretrained" in name: _lowercase = name.replace('pretrained' , 'dpt' ) if "bn" in name: _lowercase = name.replace('bn' , 'batch_norm' ) if "head" in name: _lowercase = name.replace('head' , 'head.head' ) if "encoder.norm" in name: _lowercase = name.replace('encoder.norm' , 'layernorm' ) if "auxlayer" in name: _lowercase = name.replace('auxlayer' , 'auxiliary_head.head' ) if "backbone" in name: _lowercase = name.replace('backbone' , 'backbone.bit.encoder' ) if ".." in name: _lowercase = name.replace('..' , '.' ) if "stem.conv" in name: _lowercase = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: _lowercase = name.replace('blocks' , 'layers' ) if "convolution" in name and "backbone" in name: _lowercase = name.replace('convolution' , 'conv' ) if "layer" in name and "backbone" in name: _lowercase = name.replace('layer' , 'layers' ) if "backbone.bit.encoder.bit" in name: _lowercase = name.replace('backbone.bit.encoder.bit' , 'backbone.bit' ) if "embedder.conv" in name: _lowercase = name.replace('embedder.conv' , 'embedder.convolution' ) if "backbone.bit.encoder.stem.norm" in name: _lowercase = name.replace('backbone.bit.encoder.stem.norm' , 'backbone.bit.embedder.norm' ) return name def SCREAMING_SNAKE_CASE__ ( snake_case__ :List[str] , snake_case__ :int ) -> Dict: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowercase = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) _lowercase = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _lowercase = in_proj_weight[: config.hidden_size, :] _lowercase = in_proj_bias[: config.hidden_size] _lowercase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowercase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowercase = in_proj_weight[ -config.hidden_size :, : ] _lowercase = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE__ ( ) -> Tuple: _lowercase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( snake_case__ :Optional[int] , snake_case__ :List[Any] , snake_case__ :str , snake_case__ :Any , snake_case__ :List[str] ) -> str: _lowercase , _lowercase = get_dpt_config(snake_case__ ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") _lowercase = torch.load(snake_case__ , map_location='cpu' ) # remove certain keys remove_ignore_keys_(snake_case__ ) # rename keys for key in state_dict.copy().keys(): _lowercase = state_dict.pop(snake_case__ ) _lowercase = val # read in qkv matrices read_in_q_k_v(snake_case__ , snake_case__ ) # load HuggingFace model _lowercase = DPTForSemanticSegmentation(snake_case__ ) if 'ade' in checkpoint_url else DPTForDepthEstimation(snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() # Check outputs on an image _lowercase = 480 if 'ade' in checkpoint_url else 384 _lowercase = DPTImageProcessor(size=snake_case__ ) _lowercase = prepare_img() _lowercase = image_processor(snake_case__ , return_tensors='pt' ) # forward pass _lowercase = model(**snake_case__ ).logits if 'ade' in checkpoint_url else model(**snake_case__ ).predicted_depth if show_prediction: _lowercase = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='bicubic' , align_corners=snake_case__ , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(snake_case__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: model.push_to_hub('ybelkada/dpt-hybrid-midas' ) image_processor.push_to_hub('ybelkada/dpt-hybrid-midas' ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) parser.add_argument( """--show_prediction""", action="""store_true""", ) snake_case = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
67
1
import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def SCREAMING_SNAKE_CASE__ ( snake_case__ :List[Any] , snake_case__ :Union[str, Any] , snake_case__ :Any ) -> Dict: _lowercase = OmegaConf.load(snake_case__ ) _lowercase = torch.load(snake_case__ , map_location='cpu' )['model'] _lowercase = list(state_dict.keys() ) # extract state_dict for VQVAE _lowercase = {} _lowercase = 'first_stage_model.' for key in keys: if key.startswith(snake_case__ ): _lowercase = state_dict[key] # extract state_dict for UNetLDM _lowercase = {} _lowercase = 'model.diffusion_model.' for key in keys: if key.startswith(snake_case__ ): _lowercase = state_dict[key] _lowercase = config.model.params.first_stage_config.params _lowercase = config.model.params.unet_config.params _lowercase = VQModel(**snake_case__ ).eval() vqvae.load_state_dict(snake_case__ ) _lowercase = UNetLDMModel(**snake_case__ ).eval() unet.load_state_dict(snake_case__ ) _lowercase = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule='scaled_linear' , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=snake_case__ , ) _lowercase = LDMPipeline(snake_case__ , snake_case__ , snake_case__ ) pipeline.save_pretrained(snake_case__ ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) snake_case = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) snake_case = {"""configuration_vit_mae""": ["""VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMAEConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMAEForPreTraining""", """ViTMAELayer""", """ViTMAEModel""", """ViTMAEPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """TFViTMAEForPreTraining""", """TFViTMAEModel""", """TFViTMAEPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
1
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = '''lilt''' def __init__( self : int ,__A : List[Any]=3_0522 ,__A : Tuple=768 ,__A : Tuple=12 ,__A : Union[str, Any]=12 ,__A : Optional[int]=3072 ,__A : Union[str, Any]="gelu" ,__A : str=0.1 ,__A : List[str]=0.1 ,__A : Optional[Any]=512 ,__A : List[str]=2 ,__A : Optional[int]=0.02 ,__A : Dict=1e-12 ,__A : str=0 ,__A : Tuple="absolute" ,__A : Union[str, Any]=None ,__A : Dict=4 ,__A : List[str]=1024 ,**__A : int ,) -> Any: super().__init__(pad_token_id=__A ,**__A ) _lowercase = vocab_size _lowercase = hidden_size _lowercase = num_hidden_layers _lowercase = num_attention_heads _lowercase = hidden_act _lowercase = intermediate_size _lowercase = hidden_dropout_prob _lowercase = attention_probs_dropout_prob _lowercase = max_position_embeddings _lowercase = type_vocab_size _lowercase = initializer_range _lowercase = layer_norm_eps _lowercase = position_embedding_type _lowercase = classifier_dropout _lowercase = channel_shrink_ratio _lowercase = max_ad_position_embeddings
67
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
67
1
from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( snake_case__ :Dict , snake_case__ :int , snake_case__ :List[str] , snake_case__ :int ) -> int: # noqa: E741 while r - l > 1: _lowercase = (l + r) // 2 if v[m] >= key: _lowercase = m else: _lowercase = m # noqa: E741 return r def SCREAMING_SNAKE_CASE__ ( snake_case__ :list[int] ) -> int: if len(snake_case__ ) == 0: return 0 _lowercase = [0] * len(snake_case__ ) _lowercase = 1 _lowercase = v[0] for i in range(1 , len(snake_case__ ) ): if v[i] < tail[0]: _lowercase = v[i] elif v[i] > tail[length - 1]: _lowercase = v[i] length += 1 else: _lowercase = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
67
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = '''transfo-xl''' SCREAMING_SNAKE_CASE_ : Optional[Any] = ['''mems'''] SCREAMING_SNAKE_CASE_ : Optional[Any] = { '''n_token''': '''vocab_size''', '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[Any] ,__A : Union[str, Any]=26_7735 ,__A : List[Any]=[2_0000, 4_0000, 20_0000] ,__A : Dict=1024 ,__A : str=1024 ,__A : Dict=16 ,__A : int=64 ,__A : Dict=4096 ,__A : List[Any]=4 ,__A : Optional[int]=False ,__A : Union[str, Any]=18 ,__A : Tuple=1600 ,__A : str=1000 ,__A : Dict=True ,__A : Dict=True ,__A : int=0 ,__A : Optional[int]=-1 ,__A : int=True ,__A : List[str]=0.1 ,__A : Optional[int]=0.0 ,__A : str=True ,__A : Tuple="normal" ,__A : Union[str, Any]=0.01 ,__A : Tuple=0.01 ,__A : Any=0.02 ,__A : Union[str, Any]=1e-5 ,__A : List[Any]=0 ,**__A : str ,) -> List[Any]: _lowercase = vocab_size _lowercase = [] self.cutoffs.extend(__A ) if proj_share_all_but_first: _lowercase = [False] + [True] * len(self.cutoffs ) else: _lowercase = [False] + [False] * len(self.cutoffs ) _lowercase = d_model _lowercase = d_embed _lowercase = d_head _lowercase = d_inner _lowercase = div_val _lowercase = pre_lnorm _lowercase = n_layer _lowercase = n_head _lowercase = mem_len _lowercase = same_length _lowercase = attn_type _lowercase = clamp_len _lowercase = sample_softmax _lowercase = adaptive _lowercase = dropout _lowercase = dropatt _lowercase = untie_r _lowercase = init _lowercase = init_range _lowercase = proj_init_std _lowercase = init_std _lowercase = layer_norm_epsilon super().__init__(eos_token_id=__A ,**__A ) @property def __UpperCAmelCase ( self : str ) -> Optional[int]: # Message copied from Transformer-XL documentation logger.info(F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def __UpperCAmelCase ( self : Any ,__A : Dict ) -> Optional[Any]: # Message copied from Transformer-XL documentation raise NotImplementedError( F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
67
1
import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> Tuple: random.seed(snake_case__ ) np.random.seed(snake_case__ ) torch.manual_seed(snake_case__ ) torch.cuda.manual_seed_all(snake_case__ ) # ^^ safe to call this function even if cuda is not available class A_ : """simple docstring""" def __init__( self : Dict ,__A : Iterable[torch.nn.Parameter] ,__A : float = 0.9999 ,__A : float = 0.0 ,__A : int = 0 ,__A : bool = False ,__A : Union[float, int] = 1.0 ,__A : Union[float, int] = 2 / 3 ,__A : Optional[Any] = None ,__A : Dict[str, Any] = None ,**__A : int ,) -> Tuple: if isinstance(__A ,torch.nn.Module ): _lowercase = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage`' ,'1.0.0' ,__A ,standard_warn=__A ,) _lowercase = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _lowercase = True if kwargs.get('max_value' ,__A ) is not None: _lowercase = 'The `max_value` argument is deprecated. Please use `decay` instead.' deprecate('max_value' ,'1.0.0' ,__A ,standard_warn=__A ) _lowercase = kwargs['max_value'] if kwargs.get('min_value' ,__A ) is not None: _lowercase = 'The `min_value` argument is deprecated. Please use `min_decay` instead.' deprecate('min_value' ,'1.0.0' ,__A ,standard_warn=__A ) _lowercase = kwargs['min_value'] _lowercase = list(__A ) _lowercase = [p.clone().detach() for p in parameters] if kwargs.get('device' ,__A ) is not None: _lowercase = 'The `device` argument is deprecated. Please use `to` instead.' deprecate('device' ,'1.0.0' ,__A ,standard_warn=__A ) self.to(device=kwargs['device'] ) _lowercase = None _lowercase = decay _lowercase = min_decay _lowercase = update_after_step _lowercase = use_ema_warmup _lowercase = inv_gamma _lowercase = power _lowercase = 0 _lowercase = None # set in `step()` _lowercase = model_cls _lowercase = model_config @classmethod def __UpperCAmelCase ( cls : Union[str, Any] ,__A : int ,__A : List[Any] ) -> "EMAModel": _lowercase , _lowercase = model_cls.load_config(__A ,return_unused_kwargs=__A ) _lowercase = model_cls.from_pretrained(__A ) _lowercase = cls(model.parameters() ,model_cls=__A ,model_config=model.config ) ema_model.load_state_dict(__A ) return ema_model def __UpperCAmelCase ( self : Optional[Any] ,__A : Optional[Any] ) -> Union[str, Any]: if self.model_cls is None: raise ValueError('`save_pretrained` can only be used if `model_cls` was defined at __init__.' ) if self.model_config is None: raise ValueError('`save_pretrained` can only be used if `model_config` was defined at __init__.' ) _lowercase = self.model_cls.from_config(self.model_config ) _lowercase = self.state_dict() state_dict.pop('shadow_params' ,__A ) model.register_to_config(**__A ) self.copy_to(model.parameters() ) model.save_pretrained(__A ) def __UpperCAmelCase ( self : Optional[Any] ,__A : int ) -> float: _lowercase = max(0 ,optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _lowercase = 1 - (1 + step / self.inv_gamma) ** -self.power else: _lowercase = (1 + step) / (10 + step) _lowercase = min(__A ,self.decay ) # make sure decay is not smaller than min_decay _lowercase = max(__A ,self.min_decay ) return cur_decay_value @torch.no_grad() def __UpperCAmelCase ( self : Optional[int] ,__A : Iterable[torch.nn.Parameter] ) -> Optional[Any]: if isinstance(__A ,torch.nn.Module ): _lowercase = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage.step`' ,'1.0.0' ,__A ,standard_warn=__A ,) _lowercase = parameters.parameters() _lowercase = list(__A ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _lowercase = self.get_decay(self.optimization_step ) _lowercase = decay _lowercase = 1 - decay _lowercase = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params ,__A ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _lowercase = deepspeed.zero.GatheredParameters(__A ,modifier_rank=__A ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(__A ) def __UpperCAmelCase ( self : Any ,__A : Iterable[torch.nn.Parameter] ) -> None: _lowercase = list(__A ) for s_param, param in zip(self.shadow_params ,__A ): param.data.copy_(s_param.to(param.device ).data ) def __UpperCAmelCase ( self : Any ,__A : Optional[int]=None ,__A : Union[str, Any]=None ) -> None: _lowercase = [ p.to(device=__A ,dtype=__A ) if p.is_floating_point() else p.to(device=__A ) for p in self.shadow_params ] def __UpperCAmelCase ( self : Optional[Any] ) -> dict: return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def __UpperCAmelCase ( self : Dict ,__A : Iterable[torch.nn.Parameter] ) -> None: _lowercase = [param.detach().cpu().clone() for param in parameters] def __UpperCAmelCase ( self : int ,__A : Iterable[torch.nn.Parameter] ) -> None: if self.temp_stored_params is None: raise RuntimeError('This ExponentialMovingAverage has no `store()`ed weights ' 'to `restore()`' ) for c_param, param in zip(self.temp_stored_params ,__A ): param.data.copy_(c_param.data ) # Better memory-wise. _lowercase = None def __UpperCAmelCase ( self : List[str] ,__A : dict ) -> None: _lowercase = copy.deepcopy(__A ) _lowercase = state_dict.get('decay' ,self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('Decay must be between 0 and 1' ) _lowercase = state_dict.get('min_decay' ,self.min_decay ) if not isinstance(self.min_decay ,__A ): raise ValueError('Invalid min_decay' ) _lowercase = state_dict.get('optimization_step' ,self.optimization_step ) if not isinstance(self.optimization_step ,__A ): raise ValueError('Invalid optimization_step' ) _lowercase = state_dict.get('update_after_step' ,self.update_after_step ) if not isinstance(self.update_after_step ,__A ): raise ValueError('Invalid update_after_step' ) _lowercase = state_dict.get('use_ema_warmup' ,self.use_ema_warmup ) if not isinstance(self.use_ema_warmup ,__A ): raise ValueError('Invalid use_ema_warmup' ) _lowercase = state_dict.get('inv_gamma' ,self.inv_gamma ) if not isinstance(self.inv_gamma ,(float, int) ): raise ValueError('Invalid inv_gamma' ) _lowercase = state_dict.get('power' ,self.power ) if not isinstance(self.power ,(float, int) ): raise ValueError('Invalid power' ) _lowercase = state_dict.get('shadow_params' ,__A ) if shadow_params is not None: _lowercase = shadow_params if not isinstance(self.shadow_params ,__A ): raise ValueError('shadow_params must be a list' ) if not all(isinstance(__A ,torch.Tensor ) for p in self.shadow_params ): raise ValueError('shadow_params must all be Tensors' )
67
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json""" ), } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = '''dpr''' def __init__( self : int ,__A : Union[str, Any]=3_0522 ,__A : Optional[int]=768 ,__A : int=12 ,__A : List[Any]=12 ,__A : Optional[Any]=3072 ,__A : Union[str, Any]="gelu" ,__A : Union[str, Any]=0.1 ,__A : List[Any]=0.1 ,__A : str=512 ,__A : List[str]=2 ,__A : Tuple=0.02 ,__A : Tuple=1e-12 ,__A : List[Any]=0 ,__A : List[str]="absolute" ,__A : int = 0 ,**__A : int ,) -> Tuple: super().__init__(pad_token_id=__A ,**__A ) _lowercase = vocab_size _lowercase = hidden_size _lowercase = num_hidden_layers _lowercase = num_attention_heads _lowercase = hidden_act _lowercase = intermediate_size _lowercase = hidden_dropout_prob _lowercase = attention_probs_dropout_prob _lowercase = max_position_embeddings _lowercase = type_vocab_size _lowercase = initializer_range _lowercase = layer_norm_eps _lowercase = projection_dim _lowercase = position_embedding_type
67
1
def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> int: # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection _lowercase = len(snake_case__ ) _lowercase = max(snake_case__ ) _lowercase = min(snake_case__ ) # create the counting array _lowercase = coll_max + 1 - coll_min _lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , snake_case__ ): _lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection _lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , snake_case__ ) ): _lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def SCREAMING_SNAKE_CASE__ ( snake_case__ :Union[str, Any] ) -> str: return "".join([chr(snake_case__ ) for i in counting_sort([ord(snake_case__ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" snake_case = input("""Enter numbers separated by a comma:\n""").strip() snake_case = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
67
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time snake_case = Lock() def SCREAMING_SNAKE_CASE__ ( snake_case__ :Optional[int] , snake_case__ :Union[str, Any] , snake_case__ :Tuple , snake_case__ :Any , snake_case__ :Dict , snake_case__ :Optional[int] , snake_case__ :List[str] ) -> Optional[Any]: global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(snake_case__ ) process_lock.release() # receive your right neighbor's value process_lock.acquire() _lowercase = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left _lowercase = min(snake_case__ , snake_case__ ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(snake_case__ ) process_lock.release() # receive your left neighbor's value process_lock.acquire() _lowercase = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right _lowercase = max(snake_case__ , snake_case__ ) # after all swaps are performed, send the values back to main result_pipe[1].send(snake_case__ ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :str ) -> Dict: _lowercase = [] _lowercase = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop _lowercase = Pipe() _lowercase = Pipe() process_array_.append( Process( target=snake_case__ , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) _lowercase = temp_rs _lowercase = temp_rr for i in range(1 , len(snake_case__ ) - 1 ): _lowercase = Pipe() _lowercase = Pipe() process_array_.append( Process( target=snake_case__ , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) _lowercase = temp_rs _lowercase = temp_rr process_array_.append( Process( target=snake_case__ , args=( len(snake_case__ ) - 1, arr[len(snake_case__ ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(snake_case__ ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(snake_case__ ) ): _lowercase = result_pipe[p][0].recv() process_array_[p].join() return arr def SCREAMING_SNAKE_CASE__ ( ) -> List[Any]: _lowercase = list(range(10 , 0 , -1 ) ) print('Initial List' ) print(*snake_case__ ) _lowercase = odd_even_transposition(snake_case__ ) print('Sorted List\n' ) print(*snake_case__ ) if __name__ == "__main__": main()
67
1
import math def SCREAMING_SNAKE_CASE__ ( ) -> None: _lowercase = input('Enter message: ' ) _lowercase = int(input(F"""Enter key [2-{len(snake_case__ ) - 1}]: """ ) ) _lowercase = input('Encryption/Decryption [e/d]: ' ) if mode.lower().startswith('e' ): _lowercase = encrypt_message(snake_case__ , snake_case__ ) elif mode.lower().startswith('d' ): _lowercase = decrypt_message(snake_case__ , snake_case__ ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F"""Output:\n{text + "|"}""" ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :int , snake_case__ :str ) -> str: _lowercase = [''] * key for col in range(snake_case__ ): _lowercase = col while pointer < len(snake_case__ ): cipher_text[col] += message[pointer] pointer += key return "".join(snake_case__ ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :int , snake_case__ :str ) -> str: _lowercase = math.ceil(len(snake_case__ ) / key ) _lowercase = key _lowercase = (num_cols * num_rows) - len(snake_case__ ) _lowercase = [''] * num_cols _lowercase = 0 _lowercase = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _lowercase = 0 row += 1 return "".join(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
67
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = '''big_bird''' def __init__( self : str ,__A : Union[str, Any]=5_0358 ,__A : Any=768 ,__A : List[str]=12 ,__A : Union[str, Any]=12 ,__A : int=3072 ,__A : Tuple="gelu_new" ,__A : Any=0.1 ,__A : Optional[Any]=0.1 ,__A : Tuple=4096 ,__A : int=2 ,__A : Union[str, Any]=0.02 ,__A : Optional[int]=1e-12 ,__A : List[str]=True ,__A : List[Any]=0 ,__A : Optional[Any]=1 ,__A : Optional[int]=2 ,__A : Optional[int]=66 ,__A : Tuple="block_sparse" ,__A : Optional[int]=True ,__A : Optional[int]=False ,__A : Tuple=64 ,__A : str=3 ,__A : Optional[int]=None ,**__A : Dict ,) -> Union[str, Any]: super().__init__( pad_token_id=__A ,bos_token_id=__A ,eos_token_id=__A ,sep_token_id=__A ,**__A ,) _lowercase = vocab_size _lowercase = max_position_embeddings _lowercase = hidden_size _lowercase = num_hidden_layers _lowercase = num_attention_heads _lowercase = intermediate_size _lowercase = hidden_act _lowercase = hidden_dropout_prob _lowercase = attention_probs_dropout_prob _lowercase = initializer_range _lowercase = type_vocab_size _lowercase = layer_norm_eps _lowercase = use_cache _lowercase = rescale_embeddings _lowercase = attention_type _lowercase = use_bias _lowercase = block_size _lowercase = num_random_blocks _lowercase = classifier_dropout class A_ ( UpperCAmelCase ): """simple docstring""" @property def __UpperCAmelCase ( self : int ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _lowercase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _lowercase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
67
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) snake_case = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
def SCREAMING_SNAKE_CASE__ ( snake_case__ :str ) -> list: _lowercase = [0] * len(snake_case__ ) for i in range(1 , len(snake_case__ ) ): # use last results for better performance - dynamic programming _lowercase = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: _lowercase = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 _lowercase = j return prefix_result def SCREAMING_SNAKE_CASE__ ( snake_case__ :str ) -> int: return max(prefix_function(snake_case__ ) ) if __name__ == "__main__": import doctest doctest.testmod()
67
1
def SCREAMING_SNAKE_CASE__ ( snake_case__ :Dict=2_8123 ) -> Optional[Any]: _lowercase = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i _lowercase = set() _lowercase = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(snake_case__ ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
67
def SCREAMING_SNAKE_CASE__ ( snake_case__ :str ) -> Union[str, Any]: _lowercase = len(snake_case__ ) _lowercase = sum(snake_case__ ) _lowercase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _lowercase = True for i in range(1 , s + 1 ): _lowercase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _lowercase = dp[i][j - 1] if arr[i - 1] <= j: _lowercase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _lowercase = s - 2 * j break return diff
67
1
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC snake_case = parse(importlib.metadata.version("""torch""")) def SCREAMING_SNAKE_CASE__ ( snake_case__ :Union[str, Version] , snake_case__ :str , snake_case__ :str ) -> int: if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) _lowercase = STR_OPERATION_TO_FUNC[operation] if isinstance(snake_case__ , snake_case__ ): _lowercase = parse(importlib.metadata.version(snake_case__ ) ) return operation(snake_case__ , parse(snake_case__ ) ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :str , snake_case__ :str ) -> Dict: return compare_versions(snake_case__ , snake_case__ , snake_case__ )
67
from manim import * class A_ ( UpperCAmelCase ): """simple docstring""" def __UpperCAmelCase ( self : str ) -> Union[str, Any]: _lowercase = Rectangle(height=0.5 ,width=0.5 ) _lowercase = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) _lowercase = Rectangle(height=0.25 ,width=0.25 ) _lowercase = [mem.copy() for i in range(6 )] _lowercase = [mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(__A ,__A ).arrange(__A ,buff=0 ) _lowercase = Text('CPU' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__A ) _lowercase = [mem.copy() for i in range(4 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = Text('GPU' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) gpu.move_to([-1, -1, 0] ) self.add(__A ) _lowercase = [mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = Text('Model' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) model.move_to([3, -1.0, 0] ) self.add(__A ) _lowercase = [] _lowercase = [] for i, rect in enumerate(__A ): _lowercase = fill.copy().set_fill(__A ,opacity=0.8 ) target.move_to(__A ) model_arr.append(__A ) _lowercase = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0.0 ).set_fill(__A ,opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__A ) self.add(*__A ,*__A ) _lowercase = [meta_mem.copy() for i in range(6 )] _lowercase = [meta_mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(__A ,__A ).arrange(__A ,buff=0 ) _lowercase = Text('Disk' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) disk.move_to([-4, -1.25, 0] ) self.add(__A ,__A ) _lowercase = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _lowercase = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" ,font_size=18 ,) key_text.move_to([-5, 2.4, 0] ) self.add(__A ,__A ) _lowercase = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" ,font_size=18 ,) blue_text.next_to(__A ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) self.add(__A ) _lowercase = MarkupText( F"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__A ) ) _lowercase = Square(0.3 ) input.set_fill(__A ,opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] ,__A ,buff=0.5 ) self.play(Write(__A ) ) input.generate_target() input.target.next_to(model_arr[0] ,direction=__A ,buff=0.02 ) self.play(MoveToTarget(__A ) ) self.play(FadeOut(__A ) ) _lowercase = Arrow(start=__A ,end=__A ,color=__A ,buff=0.5 ) a.next_to(model_arr[0].get_left() ,__A ,buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) _lowercase = MarkupText( F"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__A ,run_time=3 ) ) _lowercase = {'run_time': 1, 'fade_in': True, 'fade_out': True, 'buff': 0.02} self.play( Write(__A ) ,Circumscribe(model_arr[0] ,color=__A ,**__A ) ,Circumscribe(model_cpu_arr[0] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,) self.play(MoveToTarget(model_cpu_arr[0] ) ) _lowercase = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 ,__A ,buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) _lowercase = AnimationGroup( FadeOut(__A ,run_time=0.5 ) ,MoveToTarget(__A ,run_time=0.5 ) ,FadeIn(__A ,run_time=0.5 ) ,lag_ratio=0.2 ) self.play(__A ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: _lowercase = 0.7 self.play( Circumscribe(model_arr[i] ,**__A ) ,Circumscribe(cpu_left_col_base[i] ,**__A ) ,Circumscribe(cpu_left_col_base[i + 1] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,Circumscribe(model_arr[i + 1] ,color=__A ,**__A ) ,) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) ,MoveToTarget(model_cpu_arr[i + 1] ) ,) else: self.play( MoveToTarget(model_cpu_arr[i] ,run_time=0.7 ) ,MoveToTarget(model_cpu_arr[i + 1] ,run_time=0.7 ) ,) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() ,RIGHT + 0.02 ,buff=0.2 ) self.play( Circumscribe(model_arr[-1] ,color=__A ,**__A ) ,Circumscribe(cpu_left_col_base[-1] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,) self.play(MoveToTarget(model_cpu_arr[i] ) ) _lowercase = a_c _lowercase = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] ,RIGHT + 0.02 ,buff=0.5 ) self.play( FadeOut(__A ) ,FadeOut(__A ,run_time=0.5 ) ,) _lowercase = MarkupText(F"""Inference on a model too large for GPU memory\nis successfully completed.""" ,font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(__A ,run_time=3 ) ,MoveToTarget(__A ) ) self.wait()
67
1
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class A_ ( UpperCAmelCase ): """simple docstring""" def __UpperCAmelCase ( self : Tuple ) -> List[Any]: _lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__A ,'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__A ,'num_attention_heads' ) ) self.parent.assertTrue(hasattr(__A ,'num_encoder_blocks' ) ) class A_ : """simple docstring""" def __init__( self : List[Any] ,__A : List[str] ,__A : Tuple=13 ,__A : Tuple=64 ,__A : Union[str, Any]=3 ,__A : Dict=4 ,__A : Any=[2, 2, 2, 2] ,__A : int=[8, 4, 2, 1] ,__A : int=[16, 32, 64, 128] ,__A : List[Any]=[1, 4, 8, 16] ,__A : int=[1, 2, 4, 8] ,__A : Optional[int]=True ,__A : Any=True ,__A : int="gelu" ,__A : List[Any]=0.1 ,__A : List[str]=0.1 ,__A : List[Any]=0.02 ,__A : str=3 ,__A : List[str]=None ,) -> str: _lowercase = parent _lowercase = batch_size _lowercase = image_size _lowercase = num_channels _lowercase = num_encoder_blocks _lowercase = sr_ratios _lowercase = depths _lowercase = hidden_sizes _lowercase = downsampling_rates _lowercase = num_attention_heads _lowercase = is_training _lowercase = use_labels _lowercase = hidden_act _lowercase = hidden_dropout_prob _lowercase = attention_probs_dropout_prob _lowercase = initializer_range _lowercase = num_labels _lowercase = scope def __UpperCAmelCase ( self : Union[str, Any] ) -> str: _lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase = None if self.use_labels: _lowercase = ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) _lowercase = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : Dict ) -> Optional[int]: return SegformerConfig( image_size=self.image_size ,num_channels=self.num_channels ,num_encoder_blocks=self.num_encoder_blocks ,depths=self.depths ,hidden_sizes=self.hidden_sizes ,num_attention_heads=self.num_attention_heads ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,initializer_range=self.initializer_range ,) def __UpperCAmelCase ( self : Tuple ,__A : List[str] ,__A : Any ,__A : str ) -> Dict: _lowercase = SegformerModel(config=__A ) model.to(__A ) model.eval() _lowercase = model(__A ) _lowercase = _lowercase = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def __UpperCAmelCase ( self : Optional[Any] ,__A : int ,__A : Optional[int] ,__A : Optional[int] ) -> Union[str, Any]: _lowercase = self.num_labels _lowercase = SegformerForSemanticSegmentation(__A ) model.to(__A ) model.eval() _lowercase = model(__A ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) _lowercase = model(__A ,labels=__A ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss ,0.0 ) def __UpperCAmelCase ( self : Any ,__A : List[Any] ,__A : str ,__A : Optional[Any] ) -> Optional[int]: _lowercase = 1 _lowercase = SegformerForSemanticSegmentation(config=__A ) model.to(__A ) model.eval() _lowercase = torch.randint(0 ,1 ,(self.batch_size, self.image_size, self.image_size) ).to(__A ) _lowercase = model(__A ,labels=__A ) self.parent.assertGreater(result.loss ,0.0 ) def __UpperCAmelCase ( self : str ) -> int: _lowercase = self.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase = config_and_inputs _lowercase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A_ ( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ : str = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = True SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : List[Any] = False SCREAMING_SNAKE_CASE_ : List[Any] = False def __UpperCAmelCase ( self : List[Any] ) -> List[Any]: _lowercase = SegformerModelTester(self ) _lowercase = SegformerConfigTester(self ,config_class=__A ) def __UpperCAmelCase ( self : str ) -> Dict: self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Dict ) -> Optional[int]: _lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def __UpperCAmelCase ( self : Optional[int] ) -> Optional[int]: _lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*__A ) def __UpperCAmelCase ( self : List[str] ) -> Any: _lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*__A ) @unittest.skip('SegFormer does not use inputs_embeds' ) def __UpperCAmelCase ( self : Optional[Any] ) -> Dict: pass @unittest.skip('SegFormer does not have get_input_embeddings method and get_output_embeddings methods' ) def __UpperCAmelCase ( self : List[Any] ) -> Tuple: pass def __UpperCAmelCase ( self : Dict ) -> int: _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase = model_class(__A ) _lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase = [*signature.parameters.keys()] _lowercase = ['pixel_values'] self.assertListEqual(arg_names[:1] ,__A ) def __UpperCAmelCase ( self : Dict ) -> List[str]: _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() _lowercase = True for model_class in self.all_model_classes: _lowercase = True _lowercase = False _lowercase = True _lowercase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): _lowercase = model(**self._prepare_for_class(__A ,__A ) ) _lowercase = outputs.attentions _lowercase = sum(self.model_tester.depths ) self.assertEqual(len(__A ) ,__A ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowercase = True _lowercase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): _lowercase = model(**self._prepare_for_class(__A ,__A ) ) _lowercase = outputs.attentions self.assertEqual(len(__A ) ,__A ) # verify the first attentions (first block, first layer) _lowercase = (self.model_tester.image_size // 4) ** 2 _lowercase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] ,) # verify the last attentions (last block, last layer) _lowercase = (self.model_tester.image_size // 32) ** 2 _lowercase = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) ,[self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] ,) _lowercase = len(__A ) # Check attention is always last and order is fine _lowercase = True _lowercase = True _lowercase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): _lowercase = model(**self._prepare_for_class(__A ,__A ) ) self.assertEqual(out_len + 1 ,len(__A ) ) _lowercase = outputs.attentions self.assertEqual(len(__A ) ,__A ) # verify the first attentions (first block, first layer) _lowercase = (self.model_tester.image_size // 4) ** 2 _lowercase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] ,) def __UpperCAmelCase ( self : Dict ) -> int: def check_hidden_states_output(__A : Union[str, Any] ,__A : int ,__A : int ): _lowercase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): _lowercase = model(**self._prepare_for_class(__A ,__A ) ) _lowercase = outputs.hidden_states _lowercase = self.model_tester.num_encoder_blocks self.assertEqual(len(__A ) ,__A ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase = True check_hidden_states_output(__A ,__A ,__A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowercase = True check_hidden_states_output(__A ,__A ,__A ) def __UpperCAmelCase ( self : Any ) -> Optional[Any]: if not self.model_tester.is_training: return _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() _lowercase = True for model_class in self.all_model_classes: if model_class in get_values(__A ): continue _lowercase = model_class(__A ) model.to(__A ) model.train() _lowercase = self._prepare_for_class(__A ,__A ,return_labels=__A ) _lowercase = model(**__A ).loss loss.backward() @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __UpperCAmelCase ( self : Optional[Any] ) -> Tuple: pass @slow def __UpperCAmelCase ( self : Tuple ) -> Any: for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase = SegformerModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _lowercase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class A_ ( unittest.TestCase ): """simple docstring""" @slow def __UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: # only resize + normalize _lowercase = SegformerImageProcessor( image_scale=(512, 512) ,keep_ratio=__A ,align=__A ,do_random_crop=__A ) _lowercase = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( __A ) _lowercase = prepare_img() _lowercase = image_processor(images=__A ,return_tensors='pt' ) _lowercase = encoded_inputs.pixel_values.to(__A ) with torch.no_grad(): _lowercase = model(__A ) _lowercase = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape ,__A ) _lowercase = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] ,__A ,atol=1e-4 ) ) @slow def __UpperCAmelCase ( self : str ) -> Any: # only resize + normalize _lowercase = SegformerImageProcessor( image_scale=(512, 512) ,keep_ratio=__A ,align=__A ,do_random_crop=__A ) _lowercase = SegformerForSemanticSegmentation.from_pretrained( 'nvidia/segformer-b1-finetuned-cityscapes-1024-1024' ).to(__A ) _lowercase = prepare_img() _lowercase = image_processor(images=__A ,return_tensors='pt' ) _lowercase = encoded_inputs.pixel_values.to(__A ) with torch.no_grad(): _lowercase = model(__A ) _lowercase = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape ,__A ) _lowercase = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] ,__A ,atol=1e-1 ) ) @slow def __UpperCAmelCase ( self : str ) -> Any: # only resize + normalize _lowercase = SegformerImageProcessor( image_scale=(512, 512) ,keep_ratio=__A ,align=__A ,do_random_crop=__A ) _lowercase = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( __A ) _lowercase = prepare_img() _lowercase = image_processor(images=__A ,return_tensors='pt' ) _lowercase = encoded_inputs.pixel_values.to(__A ) with torch.no_grad(): _lowercase = model(__A ) _lowercase = outputs.logits.detach().cpu() _lowercase = image_processor.post_process_semantic_segmentation(outputs=__A ,target_sizes=[(500, 300)] ) _lowercase = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape ,__A ) _lowercase = image_processor.post_process_semantic_segmentation(outputs=__A ) _lowercase = torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape ,__A )
67
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class A_ : """simple docstring""" def __init__( self : Dict ,__A : Any ,__A : Tuple=None ,__A : Optional[int]=None ,__A : Optional[int]=None ,__A : int="resnet50" ,__A : int=3 ,__A : List[Any]=32 ,__A : Tuple=3 ,__A : List[Any]=True ,__A : Tuple=True ,) -> Any: _lowercase = parent _lowercase = out_indices if out_indices is not None else [4] _lowercase = stage_names _lowercase = out_features _lowercase = backbone _lowercase = batch_size _lowercase = image_size _lowercase = num_channels _lowercase = use_pretrained_backbone _lowercase = is_training def __UpperCAmelCase ( self : List[Any] ) -> Tuple: _lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase = self.get_config() return config, pixel_values def __UpperCAmelCase ( self : Tuple ) -> Tuple: return TimmBackboneConfig( image_size=self.image_size ,num_channels=self.num_channels ,out_features=self.out_features ,out_indices=self.out_indices ,stage_names=self.stage_names ,use_pretrained_backbone=self.use_pretrained_backbone ,backbone=self.backbone ,) def __UpperCAmelCase ( self : Any ,__A : Any ,__A : Dict ) -> Union[str, Any]: _lowercase = TimmBackbone(config=__A ) model.to(__A ) model.eval() with torch.no_grad(): _lowercase = model(__A ) self.parent.assertEqual( result.feature_map[-1].shape ,(self.batch_size, model.channels[-1], 14, 14) ,) def __UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: _lowercase = self.prepare_config_and_inputs() _lowercase , _lowercase = config_and_inputs _lowercase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch @require_timm class A_ ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TimmBackbone,) if is_torch_available() else () SCREAMING_SNAKE_CASE_ : List[str] = {'''feature-extraction''': TimmBackbone} if is_torch_available() else {} SCREAMING_SNAKE_CASE_ : List[Any] = False SCREAMING_SNAKE_CASE_ : Tuple = False SCREAMING_SNAKE_CASE_ : List[str] = False SCREAMING_SNAKE_CASE_ : Any = False def __UpperCAmelCase ( self : str ) -> Optional[int]: _lowercase = TimmBackboneModelTester(self ) _lowercase = ConfigTester(self ,config_class=__A ,has_text_modality=__A ) def __UpperCAmelCase ( self : int ) -> Tuple: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __UpperCAmelCase ( self : List[Any] ) -> List[str]: _lowercase = 'resnet18' _lowercase = 'microsoft/resnet-18' _lowercase = AutoBackbone.from_pretrained(__A ,use_timm_backbone=__A ) _lowercase = AutoBackbone.from_pretrained(__A ) self.assertEqual(len(timm_model.out_features ) ,len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) ,len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels ,transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices ,(-1,) ) self.assertEqual(transformers_model.out_indices ,[len(timm_model.stage_names ) - 1] ) _lowercase = AutoBackbone.from_pretrained(__A ,use_timm_backbone=__A ,out_indices=[1, 2, 3] ) _lowercase = AutoBackbone.from_pretrained(__A ,out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices ,transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) ,len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels ,transformers_model.channels ) @unittest.skip('TimmBackbone doesn\'t support feed forward chunking' ) def __UpperCAmelCase ( self : Optional[int] ) -> Any: pass @unittest.skip('TimmBackbone doesn\'t have num_hidden_layers attribute' ) def __UpperCAmelCase ( self : Optional[int] ) -> Any: pass @unittest.skip('TimmBackbone initialization is managed on the timm side' ) def __UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass @unittest.skip('TimmBackbone models doesn\'t have inputs_embeds' ) def __UpperCAmelCase ( self : Dict ) -> List[str]: pass @unittest.skip('TimmBackbone models doesn\'t have inputs_embeds' ) def __UpperCAmelCase ( self : Optional[int] ) -> Dict: pass @unittest.skip('TimmBackbone model cannot be created without specifying a backbone checkpoint' ) def __UpperCAmelCase ( self : Dict ) -> List[str]: pass @unittest.skip('Only checkpoints on timm can be loaded into TimmBackbone' ) def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: pass @unittest.skip('model weights aren\'t tied in TimmBackbone.' ) def __UpperCAmelCase ( self : Optional[Any] ) -> Optional[Any]: pass @unittest.skip('model weights aren\'t tied in TimmBackbone.' ) def __UpperCAmelCase ( self : Any ) -> List[Any]: pass @unittest.skip('Only checkpoints on timm can be loaded into TimmBackbone' ) def __UpperCAmelCase ( self : int ) -> Any: pass @unittest.skip('Only checkpoints on timm can be loaded into TimmBackbone' ) def __UpperCAmelCase ( self : List[Any] ) -> Union[str, Any]: pass @unittest.skip('TimmBackbone doesn\'t have hidden size info in its configuration.' ) def __UpperCAmelCase ( self : Any ) -> Optional[Any]: pass @unittest.skip('TimmBackbone doesn\'t support output_attentions.' ) def __UpperCAmelCase ( self : Any ) -> Any: pass @unittest.skip('Safetensors is not supported by timm.' ) def __UpperCAmelCase ( self : Union[str, Any] ) -> Tuple: pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __UpperCAmelCase ( self : int ) -> Optional[Any]: pass def __UpperCAmelCase ( self : Dict ) -> int: _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase = model_class(__A ) _lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase = [*signature.parameters.keys()] _lowercase = ['pixel_values'] self.assertListEqual(arg_names[:1] ,__A ) def __UpperCAmelCase ( self : Tuple ) -> List[str]: _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() _lowercase = True _lowercase = self.has_attentions # no need to test all models as different heads yield the same functionality _lowercase = self.all_model_classes[0] _lowercase = model_class(__A ) model.to(__A ) _lowercase = self._prepare_for_class(__A ,__A ) _lowercase = model(**__A ) _lowercase = outputs[0][-1] # Encoder-/Decoder-only models _lowercase = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: _lowercase = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=__A ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def __UpperCAmelCase ( self : List[str] ) -> int: _lowercase , _lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase = model_class(__A ) model.to(__A ) model.eval() _lowercase = model(**__A ) self.assertEqual(len(result.feature_maps ) ,len(config.out_indices ) ) self.assertEqual(len(model.channels ) ,len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None _lowercase = copy.deepcopy(__A ) _lowercase = None _lowercase = model_class(__A ) model.to(__A ) model.eval() _lowercase = model(**__A ) self.assertEqual(len(result.feature_maps ) ,1 ) self.assertEqual(len(model.channels ) ,1 ) # Check backbone can be initialized with fresh weights _lowercase = copy.deepcopy(__A ) _lowercase = False _lowercase = model_class(__A ) model.to(__A ) model.eval() _lowercase = model(**__A )
67
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def SCREAMING_SNAKE_CASE__ ( snake_case__ :BertModel , snake_case__ :str , snake_case__ :str ) -> int: _lowercase = ('dense.weight', 'attention.self.query', 'attention.self.key', 'attention.self.value') _lowercase = ( ('layer.', 'layer_'), ('word_embeddings.weight', 'word_embeddings'), ('position_embeddings.weight', 'position_embeddings'), ('token_type_embeddings.weight', 'token_type_embeddings'), ('.', '/'), ('LayerNorm/weight', 'LayerNorm/gamma'), ('LayerNorm/bias', 'LayerNorm/beta'), ('weight', 'kernel'), ) if not os.path.isdir(snake_case__ ): os.makedirs(snake_case__ ) _lowercase = model.state_dict() def to_tf_var_name(snake_case__ :str ): for patt, repl in iter(snake_case__ ): _lowercase = name.replace(snake_case__ , snake_case__ ) return F"""bert/{name}""" def create_tf_var(snake_case__ :np.ndarray , snake_case__ :str , snake_case__ :tf.Session ): _lowercase = tf.dtypes.as_dtype(tensor.dtype ) _lowercase = tf.get_variable(dtype=snake_case__ , shape=tensor.shape , name=snake_case__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(snake_case__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: _lowercase = to_tf_var_name(snake_case__ ) _lowercase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): _lowercase = torch_tensor.T _lowercase = create_tf_var(tensor=snake_case__ , name=snake_case__ , session=snake_case__ ) tf.keras.backend.set_value(snake_case__ , snake_case__ ) _lowercase = session.run(snake_case__ ) print(F"""Successfully created {tf_name}: {np.allclose(snake_case__ , snake_case__ )}""" ) _lowercase = tf.train.Saver(tf.trainable_variables() ) saver.save(snake_case__ , os.path.join(snake_case__ , model_name.replace('-' , '_' ) + '.ckpt' ) ) def SCREAMING_SNAKE_CASE__ ( snake_case__ :Any=None ) -> List[str]: _lowercase = argparse.ArgumentParser() parser.add_argument('--model_name' , type=snake_case__ , required=snake_case__ , help='model name e.g. bert-base-uncased' ) parser.add_argument( '--cache_dir' , type=snake_case__ , default=snake_case__ , required=snake_case__ , help='Directory containing pytorch model' ) parser.add_argument('--pytorch_model_path' , type=snake_case__ , required=snake_case__ , help='/path/to/<pytorch-model-name>.bin' ) parser.add_argument('--tf_cache_dir' , type=snake_case__ , required=snake_case__ , help='Directory in which to save tensorflow model' ) _lowercase = parser.parse_args(snake_case__ ) _lowercase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=snake_case__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
67
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case = { """configuration_nllb_moe""": [ """NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NllbMoeConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST""", """NllbMoeForConditionalGeneration""", """NllbMoeModel""", """NllbMoePreTrainedModel""", """NllbMoeTop2Router""", """NllbMoeSparseMLP""", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
1
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class A_ ( UpperCAmelCase , UpperCAmelCase ): """simple docstring""" @register_to_config def __init__( self : List[str] ,__A : int = 768 ,) -> List[Any]: super().__init__() _lowercase = nn.Parameter(torch.zeros(1 ,__A ) ) _lowercase = nn.Parameter(torch.ones(1 ,__A ) ) def __UpperCAmelCase ( self : List[str] ,__A : Optional[Union[str, torch.device]] = None ,__A : Optional[torch.dtype] = None ,) -> Optional[int]: _lowercase = nn.Parameter(self.mean.to(__A ).to(__A ) ) _lowercase = nn.Parameter(self.std.to(__A ).to(__A ) ) return self def __UpperCAmelCase ( self : int ,__A : str ) -> List[Any]: _lowercase = (embeds - self.mean) * 1.0 / self.std return embeds def __UpperCAmelCase ( self : str ,__A : str ) -> Tuple: _lowercase = (embeds * self.std) + self.mean return embeds
67
snake_case = {str(digit): digit**5 for digit in range(1_0)} def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(snake_case__ ) ) def SCREAMING_SNAKE_CASE__ ( ) -> int: return sum( number for number in range(1000 , 100_0000 ) if number == digits_fifth_powers_sum(snake_case__ ) ) if __name__ == "__main__": print(solution())
67
1
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name class A_ ( UpperCAmelCase ): """simple docstring""" def __init__( self : Union[str, Any] ,__A : Union[str, Any] ,__A : Optional[Any]=768 ) -> Optional[int]: super().__init__(__A ) _lowercase = proj_size _lowercase = CLIPVisionModel(__A ) _lowercase = PaintByExampleMapper(__A ) _lowercase = nn.LayerNorm(config.hidden_size ) _lowercase = nn.Linear(config.hidden_size ,self.proj_size ) # uncondition for scaling _lowercase = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def __UpperCAmelCase ( self : str ,__A : Optional[int] ,__A : Optional[int]=False ) -> Union[str, Any]: _lowercase = self.model(pixel_values=__A ) _lowercase = clip_output.pooler_output _lowercase = self.mapper(latent_states[:, None] ) _lowercase = self.final_layer_norm(__A ) _lowercase = self.proj_out(__A ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class A_ ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] ,__A : Dict ) -> str: super().__init__() _lowercase = (config.num_hidden_layers + 1) // 5 _lowercase = config.hidden_size _lowercase = 1 _lowercase = nn.ModuleList( [ BasicTransformerBlock(__A ,__A ,__A ,activation_fn='gelu' ,attention_bias=__A ) for _ in range(__A ) ] ) def __UpperCAmelCase ( self : Tuple ,__A : Optional[Any] ) -> Dict: for block in self.blocks: _lowercase = block(__A ) return hidden_states
67
from __future__ import annotations import math def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> list[int]: _lowercase = str(snake_case__ ) _lowercase = [n] for i in range(1 , len(snake_case__ ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> bool: if len(str(snake_case__ ) ) > 3: if not is_prime(int(str(snake_case__ )[-3:] ) ) or not is_prime(int(str(snake_case__ )[:3] ) ): return False return True def SCREAMING_SNAKE_CASE__ ( snake_case__ :int = 11 ) -> list[int]: _lowercase = [] _lowercase = 13 while len(snake_case__ ) != count: if validate(snake_case__ ): _lowercase = list_truncated_nums(snake_case__ ) if all(is_prime(snake_case__ ) for i in list_nums ): list_truncated_primes.append(snake_case__ ) num += 2 return list_truncated_primes def SCREAMING_SNAKE_CASE__ ( ) -> int: return sum(compute_truncated_primes(11 ) ) if __name__ == "__main__": print(F"""{sum(compute_truncated_primes(1_1)) = }""")
67
1