code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
from collections import defaultdict def lowerCAmelCase_ ( _lowercase : List[str]) -> int: """simple docstring""" a__ : Optional[int] = 1 a__ : int = True for v in tree[start]: if v not in visited: ret += dfs(_lowercase) if ret % 2 == 0: cuts.append(_lowercase) return ret def lowerCAmelCase_ ( ) -> List[str]: """simple docstring""" dfs(1) if __name__ == "__main__": _lowercase , _lowercase : List[Any] =10, 9 _lowercase : Union[str, Any] =defaultdict(list) _lowercase : Tuple ={} _lowercase : Optional[int] =[] _lowercase : Any =0 _lowercase : Tuple =[(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
170
'''simple docstring''' import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html _snake_case = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class a__ : _SCREAMING_SNAKE_CASE : Tuple = PegasusConfig _SCREAMING_SNAKE_CASE : int = {} _SCREAMING_SNAKE_CASE : Optional[int] = 'gelu' def __init__( self , _UpperCamelCase , _UpperCamelCase=13 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=False , _UpperCamelCase=99 , _UpperCamelCase=32 , _UpperCamelCase=5 , _UpperCamelCase=4 , _UpperCamelCase=37 , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=20 , _UpperCamelCase=2 , _UpperCamelCase=1 , _UpperCamelCase=0 , ): """simple docstring""" _lowercase : Union[str, Any] = parent _lowercase : Optional[int] = batch_size _lowercase : List[str] = seq_length _lowercase : Optional[Any] = is_training _lowercase : int = use_labels _lowercase : Optional[int] = vocab_size _lowercase : str = hidden_size _lowercase : Optional[Any] = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Tuple = intermediate_size _lowercase : Optional[int] = hidden_dropout_prob _lowercase : Optional[Any] = attention_probs_dropout_prob _lowercase : Tuple = max_position_embeddings _lowercase : List[str] = eos_token_id _lowercase : Optional[int] = pad_token_id _lowercase : Optional[Any] = bos_token_id def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ).clip(3 , self.vocab_size ) _lowercase : Tuple = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) , 1 ) _lowercase : int = np.concatenate([input_ids, eos_tensor] , axis=1 ) _lowercase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _lowercase : Optional[int] = prepare_pegasus_inputs_dict(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) return config, inputs_dict def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : List[Any] = 20 _lowercase : int = model_class_name(_UpperCamelCase ) _lowercase : str = model.encode(inputs_dict["input_ids"] ) _lowercase , _lowercase : List[str] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _lowercase : List[Any] = model.init_cache(decoder_input_ids.shape[0] , _UpperCamelCase , _UpperCamelCase ) _lowercase : str = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) _lowercase : Tuple = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _lowercase : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] , _UpperCamelCase , decoder_attention_mask=_UpperCamelCase , past_key_values=_UpperCamelCase , decoder_position_ids=_UpperCamelCase , ) _lowercase : Union[str, Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _lowercase : Optional[Any] = model.decode( decoder_input_ids[:, -1:] , _UpperCamelCase , decoder_attention_mask=_UpperCamelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=_UpperCamelCase , ) _lowercase : Any = model.decode(_UpperCamelCase , _UpperCamelCase ) _lowercase : int = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : List[str] = 20 _lowercase : List[Any] = model_class_name(_UpperCamelCase ) _lowercase : Dict = model.encode(inputs_dict["input_ids"] ) _lowercase , _lowercase : int = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _lowercase : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _lowercase : int = model.init_cache(decoder_input_ids.shape[0] , _UpperCamelCase , _UpperCamelCase ) _lowercase : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _lowercase : Optional[int] = model.decode( decoder_input_ids[:, :-1] , _UpperCamelCase , decoder_attention_mask=_UpperCamelCase , past_key_values=_UpperCamelCase , decoder_position_ids=_UpperCamelCase , ) _lowercase : List[Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _lowercase : Optional[int] = model.decode( decoder_input_ids[:, -1:] , _UpperCamelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=_UpperCamelCase , decoder_position_ids=_UpperCamelCase , ) _lowercase : Dict = model.decode(_UpperCamelCase , _UpperCamelCase , decoder_attention_mask=_UpperCamelCase ) _lowercase : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) def _A ( snake_case , snake_case , snake_case , snake_case=None , snake_case=None , ) -> int: if attention_mask is None: _lowercase : int = np.not_equal(snake_case , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: _lowercase : Dict = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class a__ ( lowerCamelCase_ , unittest.TestCase ): _SCREAMING_SNAKE_CASE : List[str] = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) _SCREAMING_SNAKE_CASE : List[str] = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () _SCREAMING_SNAKE_CASE : List[str] = True _SCREAMING_SNAKE_CASE : int = False _SCREAMING_SNAKE_CASE : List[Any] = False _SCREAMING_SNAKE_CASE : Optional[int] = False def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Optional[Any] = FlaxPegasusModelTester(self ) _lowercase : Union[str, Any] = ConfigTester(self , config_class=_UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def _lowerCamelCase ( self ): """simple docstring""" _lowercase , _lowercase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase , _lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase , _lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowercase : Optional[int] = self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) _lowercase : Union[str, Any] = model_class(_UpperCamelCase ) @jax.jit def encode_jitted(_UpperCamelCase , _UpperCamelCase=None , **_UpperCamelCase ): return model.encode(input_ids=_UpperCamelCase , attention_mask=_UpperCamelCase ) with self.subTest("JIT Enabled" ): _lowercase : Optional[Any] = encode_jitted(**_UpperCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _lowercase : Dict = encode_jitted(**_UpperCamelCase ).to_tuple() self.assertEqual(len(_UpperCamelCase ) , len(_UpperCamelCase ) ) for jitted_output, output in zip(_UpperCamelCase , _UpperCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase , _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowercase : Optional[int] = model_class(_UpperCamelCase ) _lowercase : Tuple = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) _lowercase : List[Any] = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): return model.decode( decoder_input_ids=_UpperCamelCase , decoder_attention_mask=_UpperCamelCase , encoder_outputs=_UpperCamelCase , ) with self.subTest("JIT Enabled" ): _lowercase : List[Any] = decode_jitted(**_UpperCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _lowercase : Optional[Any] = decode_jitted(**_UpperCamelCase ).to_tuple() self.assertEqual(len(_UpperCamelCase ) , len(_UpperCamelCase ) ) for jitted_output, output in zip(_UpperCamelCase , _UpperCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def _lowerCamelCase ( self ): """simple docstring""" for model_class_name in self.all_model_classes: _lowercase : List[Any] = model_class_name.from_pretrained("google/pegasus-large" , from_pt=_UpperCamelCase ) _lowercase : Optional[int] = np.ones((1, 1) ) _lowercase : int = model(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) @slow def _lowerCamelCase ( self ): """simple docstring""" _lowercase : str = FlaxPegasusForConditionalGeneration.from_pretrained("google/pegasus-xsum" ) _lowercase : str = PegasusTokenizer.from_pretrained("google/pegasus-xsum" ) _lowercase : str = [ " PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.", " The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" ", ] _lowercase : Optional[int] = [ "California's largest electricity provider has turned off power to hundreds of thousands of customers.", "Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.", ] _lowercase : Optional[int] = tokenizer(_UpperCamelCase , return_tensors="np" , truncation=_UpperCamelCase , max_length=512 , padding=_UpperCamelCase ) _lowercase : Tuple = model.generate(**_UpperCamelCase , num_beams=2 ).sequences _lowercase : Union[str, Any] = tokenizer.batch_decode(_UpperCamelCase , skip_special_tokens=_UpperCamelCase ) assert tgt_text == decoded
250
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class _SCREAMING_SNAKE_CASE : UpperCAmelCase_ :Tuple = LEDConfig UpperCAmelCase_ :int = {} UpperCAmelCase_ :Union[str, Any] = "gelu" def __init__( self , __A , __A=13 , __A=7 , __A=True , __A=False , __A=99 , __A=32 , __A=2 , __A=4 , __A=37 , __A=0.1 , __A=0.1 , __A=20 , __A=2 , __A=1 , __A=0 , __A=4 , ) -> Union[str, Any]: lowerCAmelCase_ :Dict = parent lowerCAmelCase_ :str = batch_size lowerCAmelCase_ :Optional[Any] = seq_length lowerCAmelCase_ :str = is_training lowerCAmelCase_ :Tuple = use_labels lowerCAmelCase_ :Tuple = vocab_size lowerCAmelCase_ :Union[str, Any] = hidden_size lowerCAmelCase_ :Optional[Any] = num_hidden_layers lowerCAmelCase_ :List[Any] = num_attention_heads lowerCAmelCase_ :List[Any] = intermediate_size lowerCAmelCase_ :List[Any] = hidden_dropout_prob lowerCAmelCase_ :Optional[int] = attention_probs_dropout_prob lowerCAmelCase_ :Optional[int] = max_position_embeddings lowerCAmelCase_ :List[str] = eos_token_id lowerCAmelCase_ :Optional[Any] = pad_token_id lowerCAmelCase_ :int = bos_token_id lowerCAmelCase_ :List[Any] = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after lowerCAmelCase_ :Tuple = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests lowerCAmelCase_ :Optional[int] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __lowerCAmelCase ( self ) -> Union[str, Any]: lowerCAmelCase_ :Dict = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCAmelCase_ :Optional[int] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCAmelCase_ :Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCAmelCase_ :Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ :Optional[int] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) lowerCAmelCase_ :Union[str, Any] = prepare_led_inputs_dict(__A , __A , __A ) lowerCAmelCase_ :Tuple = tf.concat( [tf.zeros_like(__A )[:, :-1], tf.ones_like(__A )[:, -1:]] , axis=-1 , ) lowerCAmelCase_ :str = global_attention_mask return config, inputs_dict def __lowerCAmelCase ( self , __A , __A ) -> str: lowerCAmelCase_ :List[str] = TFLEDModel(config=__A ).get_decoder() lowerCAmelCase_ :Optional[Any] = inputs_dict["""input_ids"""] lowerCAmelCase_ :List[str] = input_ids[:1, :] lowerCAmelCase_ :Optional[Any] = inputs_dict["""attention_mask"""][:1, :] lowerCAmelCase_ :Any = 1 # first forward pass lowerCAmelCase_ :List[str] = model(__A , attention_mask=__A , use_cache=__A ) lowerCAmelCase_ , lowerCAmelCase_ :Any = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase_ :List[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase_ :Any = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCAmelCase_ :Dict = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCAmelCase_ :Union[str, Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCAmelCase_ :Dict = model(__A , attention_mask=__A )[0] lowerCAmelCase_ :int = model(__A , attention_mask=__A , past_key_values=__A )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCAmelCase_ :str = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCAmelCase_ :Tuple = output_from_no_past[:, -3:, random_slice_idx] lowerCAmelCase_ :Any = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__A , __A , rtol=1E-3 ) def _snake_case ( lowercase__ : int , lowercase__ : List[Any] , lowercase__ : Tuple , lowercase__ : Any=None , lowercase__ : int=None , lowercase__ : str=None , lowercase__ : Optional[int]=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowerCAmelCase_ :Union[str, Any] = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase_ :int = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase_ :Any = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase_ :Optional[int] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class _SCREAMING_SNAKE_CASE ( A__ , A__ , unittest.TestCase ): UpperCAmelCase_ :Union[str, Any] = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () UpperCAmelCase_ :Tuple = (TFLEDForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase_ :Optional[int] = ( { "conversational": TFLEDForConditionalGeneration, "feature-extraction": TFLEDModel, "summarization": TFLEDForConditionalGeneration, "text2text-generation": TFLEDForConditionalGeneration, "translation": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase_ :int = True UpperCAmelCase_ :Optional[Any] = False UpperCAmelCase_ :Union[str, Any] = False UpperCAmelCase_ :Optional[Any] = False def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :Optional[Any] = TFLEDModelTester(self ) lowerCAmelCase_ :List[str] = ConfigTester(self , config_class=__A ) def __lowerCAmelCase ( self ) -> str: self.config_tester.run_common_tests() def __lowerCAmelCase ( self ) -> str: lowerCAmelCase_ :Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__A ) def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ , lowerCAmelCase_ :Any = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ :Optional[int] = tf.zeros_like(inputs_dict["""attention_mask"""] ) lowerCAmelCase_ :Optional[int] = 2 lowerCAmelCase_ :Any = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["""global_attention_mask"""] , ) lowerCAmelCase_ :List[str] = True lowerCAmelCase_ :Optional[Any] = self.model_tester.seq_length lowerCAmelCase_ :Optional[Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(__A ): lowerCAmelCase_ :str = outputs.decoder_attentions self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(__A ): lowerCAmelCase_ :Any = [t.numpy() for t in outputs.encoder_attentions] lowerCAmelCase_ :str = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: lowerCAmelCase_ :Dict = True lowerCAmelCase_ :List[str] = False lowerCAmelCase_ :int = False lowerCAmelCase_ :Any = model_class(__A ) lowerCAmelCase_ :Any = model(self._prepare_for_class(__A , __A ) ) lowerCAmelCase_ :Union[str, Any] = len(__A ) self.assertEqual(config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) if self.is_encoder_decoder: lowerCAmelCase_ :Any = model_class(__A ) lowerCAmelCase_ :str = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(config.output_hidden_states , __A ) check_decoder_attentions_output(__A ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowerCAmelCase_ :Any = True lowerCAmelCase_ :Dict = model_class(__A ) lowerCAmelCase_ :int = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) # Check attention is always last and order is fine lowerCAmelCase_ :Dict = True lowerCAmelCase_ :Optional[int] = True lowerCAmelCase_ :str = model_class(__A ) lowerCAmelCase_ :List[Any] = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__A ) ) self.assertEqual(model.config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) @unittest.skip("""LED keeps using potentially symbolic tensors in conditionals and breaks tracing.""" ) def __lowerCAmelCase ( self ) -> str: pass def __lowerCAmelCase ( self ) -> List[str]: # TODO: Head-masking not yet implement pass def _snake_case ( lowercase__ : List[Any] ) -> int: '''simple docstring''' return tf.constant(lowercase__ , dtype=tf.intaa ) __UpperCAmelCase = 1e-4 @slow @require_tf class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :Union[str, Any] = TFLEDForConditionalGeneration.from_pretrained("""allenai/led-base-16384""" ).led # change to intended input here lowerCAmelCase_ :Union[str, Any] = _long_tensor([512 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) lowerCAmelCase_ :List[str] = _long_tensor([128 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) lowerCAmelCase_ :List[Any] = prepare_led_inputs_dict(model.config , __A , __A ) lowerCAmelCase_ :Optional[Any] = model(**__A )[0] lowerCAmelCase_ :Tuple = (1, 1024, 768) self.assertEqual(output.shape , __A ) # change to expected output here lowerCAmelCase_ :Optional[Any] = tf.convert_to_tensor( [[2.3_0_5_0, 2.8_2_7_9, 0.6_5_3_1], [-1.8_4_5_7, -0.1_4_5_5, -3.5_6_6_1], [-1.0_1_8_6, 0.4_5_8_6, -2.2_0_4_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , __A , atol=1E-3 ) def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ :List[str] = TFLEDForConditionalGeneration.from_pretrained("""allenai/led-base-16384""" ) # change to intended input here lowerCAmelCase_ :Any = _long_tensor([512 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) lowerCAmelCase_ :List[str] = _long_tensor([128 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) lowerCAmelCase_ :Tuple = prepare_led_inputs_dict(model.config , __A , __A ) lowerCAmelCase_ :Any = model(**__A )[0] lowerCAmelCase_ :Optional[Any] = (1, 1024, model.config.vocab_size) self.assertEqual(output.shape , __A ) # change to expected output here lowerCAmelCase_ :Optional[Any] = tf.convert_to_tensor( [[3_3.6_5_0_7, 6.4_5_7_2, 1_6.8_0_8_9], [5.8_7_3_9, -2.4_2_3_8, 1_1.2_9_0_2], [-3.2_1_3_9, -4.3_1_4_9, 4.2_7_8_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , __A , atol=1E-3 , rtol=1E-3 )
1
"""simple docstring""" import pandas as pd from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split # Fitting Polynomial Regression to the dataset from sklearn.preprocessing import PolynomialFeatures # Importing the dataset __UpperCAmelCase = pd.read_csv( 'https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/' 'position_salaries.csv' ) __UpperCAmelCase = dataset.iloc[:, 1:2].values __UpperCAmelCase = dataset.iloc[:, 2].values __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = train_test_split(X, y, test_size=0.2, random_state=0) __UpperCAmelCase = PolynomialFeatures(degree=4) __UpperCAmelCase = poly_reg.fit_transform(X) __UpperCAmelCase = LinearRegression() pol_reg.fit(X_poly, y) def _snake_case ( ) -> str: '''simple docstring''' plt.scatter(lowercase__ , lowercase__ , color="""red""" ) plt.plot(lowercase__ , pol_reg.predict(poly_reg.fit_transform(lowercase__ ) ) , color="""blue""" ) plt.title("""Truth or Bluff (Linear Regression)""" ) plt.xlabel("""Position level""" ) plt.ylabel("""Salary""" ) plt.show() if __name__ == "__main__": viz_polymonial() # Predicting a new result with Polymonial Regression pol_reg.predict(poly_reg.fit_transform([[5.5]])) # output should be 132148.43750003
1
1
import datasets _UpperCAmelCase : Tuple = "\\n@InProceedings{conneau2018xnli,\n author = \"Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin\",\n title = \"XNLI: Evaluating Cross-lingual Sentence Representations\",\n booktitle = \"Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n location = \"Brussels, Belgium\",\n}\n" _UpperCAmelCase : List[Any] = "\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n" _UpperCAmelCase : Any = "\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n 'accuracy': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric(\"xnli\")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n" def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class __lowerCAmelCase ( datasets.Metric): def SCREAMING_SNAKE_CASE ( self: Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Tuple ): return {"accuracy": simple_accuracy(_lowerCAmelCase , _lowerCAmelCase )}
236
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCAmelCase : List[str] = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys _UpperCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
236
1
"""simple docstring""" import argparse import copy def __SCREAMING_SNAKE_CASE ( A_ ) -> Optional[int]: lowerCAmelCase__ : Union[str, Any] = {} with open(A_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowerCAmelCase__ : Tuple = [] _list.append([line.split()[1], line.split()[2]] ) lowerCAmelCase__ : Optional[int] = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowerCAmelCase__ : Optional[Any] = [] _list.append([line.split()[0], line.split()[2]] ) lowerCAmelCase__ : Optional[Any] = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __SCREAMING_SNAKE_CASE ( A_ , A_ ) -> List[str]: with open(A_ ) as f: lowerCAmelCase__ : Any = f.read(1 ) lowerCAmelCase__ : Tuple = start_node lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : Dict = start_node lowerCAmelCase__ : Optional[int] = 0 while visiting not in first_solution: lowerCAmelCase__ : List[str] = 1_00_00 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(A_ ) and k[0] not in first_solution: lowerCAmelCase__ : Tuple = k[1] lowerCAmelCase__ : Optional[Any] = k[0] first_solution.append(A_ ) lowerCAmelCase__ : int = distance_of_first_solution + int(A_ ) lowerCAmelCase__ : Optional[int] = best_node first_solution.append(A_ ) lowerCAmelCase__ : int = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowerCAmelCase__ : Union[str, Any] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_00_00 ) return first_solution, distance_of_first_solution def __SCREAMING_SNAKE_CASE ( A_ , A_ ) -> List[Any]: lowerCAmelCase__ : Optional[int] = [] for n in solution[1:-1]: lowerCAmelCase__ : int = solution.index(A_ ) for kn in solution[1:-1]: lowerCAmelCase__ : Optional[Any] = solution.index(A_ ) if n == kn: continue lowerCAmelCase__ : Tuple = copy.deepcopy(A_ ) lowerCAmelCase__ : Tuple = kn lowerCAmelCase__ : Any = n lowerCAmelCase__ : Tuple = 0 for k in _tmp[:-1]: lowerCAmelCase__ : Optional[int] = _tmp[_tmp.index(A_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowerCAmelCase__ : Tuple = distance + int(i[1] ) _tmp.append(A_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowerCAmelCase__ : Optional[int] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda A_ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ , A_ ) -> Optional[Any]: lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : Tuple = first_solution lowerCAmelCase__ : Optional[int] = [] lowerCAmelCase__ : str = distance_of_first_solution lowerCAmelCase__ : Optional[Any] = solution while count <= iters: lowerCAmelCase__ : Any = find_neighborhood(A_ , A_ ) lowerCAmelCase__ : Union[str, Any] = 0 lowerCAmelCase__ : str = neighborhood[index_of_best_solution] lowerCAmelCase__ : Union[str, Any] = len(A_ ) - 1 lowerCAmelCase__ : Optional[int] = False while not found: lowerCAmelCase__ : Dict = 0 while i < len(A_ ): if best_solution[i] != solution[i]: lowerCAmelCase__ : int = best_solution[i] lowerCAmelCase__ : Optional[Any] = solution[i] break lowerCAmelCase__ : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowerCAmelCase__ : Optional[Any] = True lowerCAmelCase__ : List[str] = best_solution[:-1] lowerCAmelCase__ : int = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowerCAmelCase__ : str = cost lowerCAmelCase__ : int = solution else: lowerCAmelCase__ : List[str] = index_of_best_solution + 1 lowerCAmelCase__ : List[str] = neighborhood[index_of_best_solution] if len(A_ ) >= size: tabu_list.pop(0 ) lowerCAmelCase__ : Tuple = count + 1 return best_solution_ever, best_cost def __SCREAMING_SNAKE_CASE ( A_=None ) -> Any: lowerCAmelCase__ : Union[str, Any] = generate_neighbours(args.File ) lowerCAmelCase__ : List[str] = generate_first_solution( args.File , A_ ) lowerCAmelCase__ : Optional[int] = tabu_search( A_ , A_ , A_ , args.Iterations , args.Size , ) print(f'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": __UpperCamelCase : str = argparse.ArgumentParser(description='''Tabu Search''') parser.add_argument( '''-f''', '''--File''', type=str, help='''Path to the file containing the data''', required=True, ) parser.add_argument( '''-i''', '''--Iterations''', type=int, help='''How many iterations the algorithm should perform''', required=True, ) parser.add_argument( '''-s''', '''--Size''', type=int, help='''Size of the tabu list''', required=True ) # Pass the arguments to main method main(parser.parse_args())
359
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): lowerCAmelCase__ : List[str] = Rectangle(height=0.5 ,width=0.5 ) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(6 )] lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[int] = VGroup(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''CPU''' ,font_size=2_4 ) lowerCAmelCase__ : Optional[int] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(4 )] lowerCAmelCase__ : List[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Any = Text('''GPU''' ,font_size=2_4 ) lowerCAmelCase__ : str = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[str] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Tuple = Text('''Model''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 / 4 ,width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ ,opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) ,buff=0.02 ,direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] ,direction=lowercase_ ,buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] ,direction=lowercase_ ,buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) lowerCAmelCase__ : List[str] = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[int] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''Loaded Checkpoint''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,aligned_edge=lowercase_ ,buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowerCAmelCase__ : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase__ : Optional[Any] = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' ,font_size=1_8 ,) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ ,lowercase_ ) lowerCAmelCase__ : Tuple = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' ,font_size=1_8 ,) blue_text.next_to(lowercase_ ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) lowerCAmelCase__ : Tuple = MarkupText( F'Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.' ,font_size=2_4 ,) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) ,Write(lowercase_ ) ) self.play(Write(lowercase_ ,run_time=1 ) ,Create(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : int = [] lowerCAmelCase__ : Optional[Any] = [] for i, rect in enumerate(lowercase_ ): lowerCAmelCase__ : List[str] = fill.copy().set_fill(lowercase_ ,opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : List[Any] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ ,run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
74
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
177
'''simple docstring''' import math def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False UpperCAmelCase__ = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str]=1 , **SCREAMING_SNAKE_CASE__ : List[str] ): '''simple docstring''' UpperCAmelCase__ = factor * value UpperCAmelCase__ = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
346
0
'''simple docstring''' import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput a : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name def __lowerCamelCase ( _lowercase ) -> Union[str, Any]: warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , _lowercase , ) if isinstance(_lowercase , torch.Tensor ): return image elif isinstance(_lowercase , PIL.Image.Image ): UpperCAmelCase : int = [image] if isinstance(image[0] , PIL.Image.Image ): UpperCAmelCase , UpperCAmelCase : Optional[Any] = image[0].size UpperCAmelCase , UpperCAmelCase : List[Any] = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 UpperCAmelCase : Dict = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] UpperCAmelCase : str = np.concatenate(_lowercase , axis=0 ) UpperCAmelCase : str = np.array(_lowercase ).astype(np.floataa ) / 255.0 UpperCAmelCase : Union[str, Any] = image.transpose(0 , 3 , 1 , 2 ) UpperCAmelCase : List[str] = 2.0 * image - 1.0 UpperCAmelCase : List[Any] = torch.from_numpy(_lowercase ) elif isinstance(image[0] , torch.Tensor ): UpperCAmelCase : Tuple = torch.cat(_lowercase , dim=0 ) return image def __lowerCamelCase ( _lowercase ) -> List[Any]: if isinstance(_lowercase , torch.Tensor ): return mask elif isinstance(_lowercase , PIL.Image.Image ): UpperCAmelCase : Optional[Any] = [mask] if isinstance(mask[0] , PIL.Image.Image ): UpperCAmelCase , UpperCAmelCase : str = mask[0].size UpperCAmelCase , UpperCAmelCase : Optional[int] = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 UpperCAmelCase : Any = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] UpperCAmelCase : Union[str, Any] = np.concatenate(_lowercase , axis=0 ) UpperCAmelCase : List[Any] = mask.astype(np.floataa ) / 255.0 UpperCAmelCase : List[Any] = 0 UpperCAmelCase : Dict = 1 UpperCAmelCase : Dict = torch.from_numpy(_lowercase ) elif isinstance(mask[0] , torch.Tensor ): UpperCAmelCase : Union[str, Any] = torch.cat(_lowercase , dim=0 ) return mask class UpperCamelCase_ ( __magic_name__ ): lowercase = 42 lowercase = 42 def __init__( self , A , A ) -> str: super().__init__() self.register_modules(unet=A , scheduler=A ) @torch.no_grad() def __call__( self , A , A , A = 250 , A = 0.0 , A = 10 , A = 10 , A = None , A = "pil" , A = True , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : str = image UpperCAmelCase : Dict = _preprocess_image(A ) UpperCAmelCase : Optional[int] = original_image.to(device=self.device , dtype=self.unet.dtype ) UpperCAmelCase : str = _preprocess_mask(A ) UpperCAmelCase : List[str] = mask_image.to(device=self.device , dtype=self.unet.dtype ) UpperCAmelCase : Union[str, Any] = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(A , A ) and len(A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) UpperCAmelCase : str = original_image.shape UpperCAmelCase : Optional[Any] = randn_tensor(A , generator=A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(A , A , A , self.device ) UpperCAmelCase : List[str] = eta UpperCAmelCase : Optional[int] = self.scheduler.timesteps[0] + 1 UpperCAmelCase : List[Any] = generator[0] if isinstance(A , A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual UpperCAmelCase : List[Any] = self.unet(A , A ).sample # compute previous image: x_t -> x_t-1 UpperCAmelCase : str = self.scheduler.step(A , A , A , A , A , A ).prev_sample else: # compute the reverse: x_t-1 -> x_t UpperCAmelCase : Tuple = self.scheduler.undo_step(A , A , A ) UpperCAmelCase : str = t UpperCAmelCase : Optional[int] = (image / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
338
'''simple docstring''' import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING a : Union[str, Any] = logging.get_logger(__name__) a : Union[str, Any] = { """facebook/detr-resnet-50""": """https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json""", # See all DETR models at https://huggingface.co/models?filter=detr } class UpperCamelCase_ ( __magic_name__ ): lowercase = 'detr' lowercase = ['past_key_values'] lowercase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , A=True , A=None , A=3 , A=100 , A=6 , A=2048 , A=8 , A=6 , A=2048 , A=8 , A=0.0 , A=0.0 , A=True , A="relu" , A=256 , A=0.1 , A=0.0 , A=0.0 , A=0.0_2 , A=1.0 , A=False , A="sine" , A="resnet50" , A=True , A=False , A=1 , A=5 , A=2 , A=1 , A=1 , A=5 , A=2 , A=0.1 , **A , ) -> List[str]: if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(A , A ): UpperCAmelCase : Any = backbone_config.get("""model_type""" ) UpperCAmelCase : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase : List[Any] = config_class.from_dict(A ) # set timm attributes to None UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Any = None, None, None UpperCAmelCase : Dict = use_timm_backbone UpperCAmelCase : Any = backbone_config UpperCAmelCase : List[Any] = num_channels UpperCAmelCase : int = num_queries UpperCAmelCase : List[str] = d_model UpperCAmelCase : Tuple = encoder_ffn_dim UpperCAmelCase : Optional[Any] = encoder_layers UpperCAmelCase : Any = encoder_attention_heads UpperCAmelCase : Optional[Any] = decoder_ffn_dim UpperCAmelCase : Optional[int] = decoder_layers UpperCAmelCase : Any = decoder_attention_heads UpperCAmelCase : str = dropout UpperCAmelCase : Tuple = attention_dropout UpperCAmelCase : Dict = activation_dropout UpperCAmelCase : Tuple = activation_function UpperCAmelCase : List[Any] = init_std UpperCAmelCase : str = init_xavier_std UpperCAmelCase : List[Any] = encoder_layerdrop UpperCAmelCase : int = decoder_layerdrop UpperCAmelCase : List[Any] = encoder_layers UpperCAmelCase : Union[str, Any] = auxiliary_loss UpperCAmelCase : str = position_embedding_type UpperCAmelCase : Union[str, Any] = backbone UpperCAmelCase : List[str] = use_pretrained_backbone UpperCAmelCase : Optional[int] = dilation # Hungarian matcher UpperCAmelCase : Union[str, Any] = class_cost UpperCAmelCase : Optional[Any] = bbox_cost UpperCAmelCase : List[Any] = giou_cost # Loss coefficients UpperCAmelCase : int = mask_loss_coefficient UpperCAmelCase : Optional[int] = dice_loss_coefficient UpperCAmelCase : Dict = bbox_loss_coefficient UpperCAmelCase : Any = giou_loss_coefficient UpperCAmelCase : Any = eos_coefficient super().__init__(is_encoder_decoder=A , **A ) @property def _lowercase( self ) -> int: return self.encoder_attention_heads @property def _lowercase( self ) -> int: return self.d_model @classmethod def _lowercase( cls , A , **A ) -> Dict: return cls(backbone_config=A , **A ) def _lowercase( self ) -> Dict[str, any]: UpperCAmelCase : Any = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase : Any = self.backbone_config.to_dict() UpperCAmelCase : Optional[Any] = self.__class__.model_type return output class UpperCamelCase_ ( __magic_name__ ): lowercase = version.parse('1.11' ) @property def _lowercase( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _lowercase( self ) -> float: return 1e-5 @property def _lowercase( self ) -> int: return 12
338
1
"""simple docstring""" from collections import namedtuple import requests from lxml import html # type: ignore UpperCamelCase : Union[str, Any] = namedtuple("covid_data", "cases deaths recovered") def A ( snake_case :str = "https://www.worldometers.info/coronavirus/" ) -> covid_data: __UpperCamelCase = '//div[@class = "maincounter-number"]/span/text()' return covid_data(*html.fromstring(requests.get(snake_case ).content ).xpath(snake_case ) ) UpperCamelCase : List[str] = "Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}" print(fmt.format(*covid_stats()))
316
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path UpperCamelCase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) UpperCamelCase : list[int] = [ord(letter) for letter in string.ascii_lowercase] UpperCamelCase : set[int] = {ord(char) for char in VALID_CHARS} UpperCamelCase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def A ( snake_case :list[int] , snake_case :tuple[int, ...] ) -> str | None: __UpperCamelCase = "" __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 for keychar, cipherchar in zip(cycle(snake_case ) , snake_case ): __UpperCamelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(snake_case ) return decoded def A ( snake_case :list[int] ) -> list[str]: __UpperCamelCase = [] for key in product(snake_case , repeat=3 ): __UpperCamelCase = try_key(snake_case , snake_case ) if encoded is not None: possibles.append(snake_case ) return possibles def A ( snake_case :list[str] , snake_case :str ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def A ( snake_case :str = "p059_cipher.txt" ) -> int: __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = Path(snake_case ).parent.joinpath(snake_case ).read_text(encoding='utf-8' ) __UpperCamelCase = [int(snake_case ) for number in data.strip().split(',' )] __UpperCamelCase = filter_valid_chars(snake_case ) for common_word in COMMON_WORDS: __UpperCamelCase = filter_common_word(snake_case , snake_case ) if len(snake_case ) == 1: break __UpperCamelCase = possibles[0] return sum(ord(snake_case ) for char in decoded_text ) if __name__ == "__main__": print(f'''{solution() = }''')
316
1
"""simple docstring""" def __A (_SCREAMING_SNAKE_CASE ): """simple docstring""" return " ".join( ''.join(word[::-1] ) if len(snake_case__ ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
365
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __A = random.Random() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Union[str, Any]: """simple docstring""" if rng is None: lowerCAmelCase__ :int = global_rng lowerCAmelCase__ :str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=7 , __UpperCAmelCase=4_0_0 , __UpperCAmelCase=2_0_0_0 , __UpperCAmelCase=1_0 , __UpperCAmelCase=1_6_0 , __UpperCAmelCase=8 , __UpperCAmelCase=0.0 , __UpperCAmelCase=4_0_0_0 , __UpperCAmelCase=False , __UpperCAmelCase=True , ): '''simple docstring''' lowerCAmelCase__ :int = parent lowerCAmelCase__ :Optional[int] = batch_size lowerCAmelCase__ :Optional[Any] = min_seq_length lowerCAmelCase__ :Optional[int] = max_seq_length lowerCAmelCase__ :Dict = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase__ :Union[str, Any] = padding_value lowerCAmelCase__ :Optional[int] = sampling_rate lowerCAmelCase__ :Optional[int] = return_attention_mask lowerCAmelCase__ :Union[str, Any] = do_normalize lowerCAmelCase__ :Any = feature_size lowerCAmelCase__ :Union[str, Any] = chunk_length lowerCAmelCase__ :List[Any] = hop_length def snake_case ( self ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case ( self , __UpperCAmelCase=False , __UpperCAmelCase=False ): '''simple docstring''' def _flatten(__UpperCAmelCase ): return list(itertools.chain(*__UpperCAmelCase ) ) if equal_length: lowerCAmelCase__ :Any = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCAmelCase__ :Union[str, Any] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase__ :Optional[int] = [np.asarray(__UpperCAmelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Union[str, Any] = WhisperFeatureExtractor if is_speech_available() else None def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = WhisperFeatureExtractionTester(self ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :Optional[Any] = feat_extract_first.save_pretrained(__UpperCAmelCase )[0] check_json_file_has_correct_format(__UpperCAmelCase ) lowerCAmelCase__ :List[str] = self.feature_extraction_class.from_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :str = feat_extract_first.to_dict() lowerCAmelCase__ :List[Any] = feat_extract_second.to_dict() lowerCAmelCase__ :int = feat_extract_first.mel_filters lowerCAmelCase__ :Optional[int] = feat_extract_second.mel_filters self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :Optional[int] = os.path.join(__UpperCAmelCase , 'feat_extract.json' ) feat_extract_first.to_json_file(__UpperCAmelCase ) lowerCAmelCase__ :str = self.feature_extraction_class.from_json_file(__UpperCAmelCase ) lowerCAmelCase__ :Tuple = feat_extract_first.to_dict() lowerCAmelCase__ :List[Any] = feat_extract_second.to_dict() lowerCAmelCase__ :str = feat_extract_first.mel_filters lowerCAmelCase__ :Optional[int] = feat_extract_second.mel_filters self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase__ :List[Any] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] lowerCAmelCase__ :int = [np.asarray(__UpperCAmelCase ) for speech_input in speech_inputs] # Test feature size lowerCAmelCase__ :int = feature_extractor(__UpperCAmelCase , padding='max_length' , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input lowerCAmelCase__ :Tuple = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCAmelCase__ :Dict = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) # Test batched lowerCAmelCase__ :Optional[Any] = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features lowerCAmelCase__ :Dict = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(__UpperCAmelCase , __UpperCAmelCase ): self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase__ :List[str] = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] lowerCAmelCase__ :Optional[Any] = np.asarray(__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features lowerCAmelCase__ :List[Any] = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(__UpperCAmelCase , __UpperCAmelCase ): self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) # Test truncation required lowerCAmelCase__ :Any = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] lowerCAmelCase__ :Any = [np.asarray(__UpperCAmelCase ) for speech_input in speech_inputs] lowerCAmelCase__ :str = [x[: feature_extractor.n_samples] for x in speech_inputs] lowerCAmelCase__ :Union[str, Any] = [np.asarray(__UpperCAmelCase ) for speech_input in speech_inputs_truncated] lowerCAmelCase__ :Any = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features lowerCAmelCase__ :int = feature_extractor(__UpperCAmelCase , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(__UpperCAmelCase , __UpperCAmelCase ): self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) def snake_case ( self ): '''simple docstring''' import torch lowerCAmelCase__ :str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ :Dict = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) lowerCAmelCase__ :Union[str, Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase__ :List[str] = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCAmelCase__ :List[str] = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Dict = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCAmelCase__ :str = ds.sort('id' ).select(range(__UpperCAmelCase ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = torch.tensor( [ 0.11_93, -0.09_46, -0.10_98, -0.01_96, 0.02_25, -0.06_90, -0.17_36, 0.09_51, 0.09_71, -0.08_17, -0.07_02, 0.01_62, 0.02_60, 0.00_17, -0.01_92, -0.16_78, 0.07_09, -0.18_67, -0.06_55, -0.02_74, -0.02_34, -0.18_84, -0.05_16, -0.05_54, -0.02_74, -0.14_25, -0.14_23, 0.08_37, 0.03_77, -0.08_54 ] ) # fmt: on lowerCAmelCase__ :Tuple = self._load_datasamples(1 ) lowerCAmelCase__ :Any = WhisperFeatureExtractor() lowerCAmelCase__ :List[str] = feature_extractor(__UpperCAmelCase , return_tensors='pt' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , __UpperCAmelCase , atol=1E-4 ) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ :int = self._load_datasamples(1 )[0] lowerCAmelCase__ :Any = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue lowerCAmelCase__ :Any = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=__UpperCAmelCase )[0] self.assertTrue(np.all(np.mean(__UpperCAmelCase ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__UpperCAmelCase ) - 1 ) < 1E-3 ) )
254
0
'''simple docstring''' def snake_case ( UpperCAmelCase )-> List[Any]: """simple docstring""" __A , __A = [], [] while len(SCREAMING_SNAKE_CASE_ ) > 1: __A , __A = min(SCREAMING_SNAKE_CASE_ ), max(SCREAMING_SNAKE_CASE_ ) start.append(SCREAMING_SNAKE_CASE_ ) end.append(SCREAMING_SNAKE_CASE_ ) collection.remove(SCREAMING_SNAKE_CASE_ ) collection.remove(SCREAMING_SNAKE_CASE_ ) end.reverse() return start + collection + end if __name__ == "__main__": a__ : str = input("Enter numbers separated by a comma:\n").strip() a__ : List[str] = [int(item) for item in user_input.split(",")] print(*merge_sort(unsorted), sep=",")
161
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class a__ : def __init__( self , _A , _A=1_3 , _A=3_0 , _A=2 , _A=3 , _A=True , _A=True , _A=3_2 , _A=5 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=1_0 , _A=0.02 , _A=3 , _A=None , _A=2 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = scope __lowerCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) __lowerCAmelCase = (image_size // patch_size) ** 2 __lowerCAmelCase = num_patches + 2 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = self.get_config() return config, pixel_values, labels def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_A , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = DeiTModel(config=_A ) model.to(_A ) model.eval() __lowerCAmelCase = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = DeiTForMaskedImageModeling(config=_A ) model.to(_A ) model.eval() __lowerCAmelCase = model(_A ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowerCAmelCase = 1 __lowerCAmelCase = DeiTForMaskedImageModeling(_A ) model.to(_A ) model.eval() __lowerCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase = model(_A ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = self.type_sequence_label_size __lowerCAmelCase = DeiTForImageClassification(_A ) model.to(_A ) model.eval() __lowerCAmelCase = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCAmelCase = 1 __lowerCAmelCase = DeiTForImageClassification(_A ) model.to(_A ) model.eval() __lowerCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a__ ( snake_case__ , snake_case__ , unittest.TestCase ): _a : Optional[Any] = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) _a : int = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) _a : Optional[Any] = False _a : Tuple = False _a : Tuple = False def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = DeiTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=3_7 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(_A ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A=False ): """simple docstring""" __lowerCAmelCase = super()._prepare_for_class(_A , _A , return_labels=_A ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if not self.model_tester.is_training: return __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_A ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue __lowerCAmelCase = model_class(_A ) model.to(_A ) model.train() __lowerCAmelCase = self._prepare_for_class(_A , _A , return_labels=_A ) __lowerCAmelCase = model(**_A ).loss loss.backward() def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowerCAmelCase = False __lowerCAmelCase = True for model_class in self.all_model_classes: if model_class in get_values(_A ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue __lowerCAmelCase = model_class(_A ) model.gradient_checkpointing_enable() model.to(_A ) model.train() __lowerCAmelCase = self._prepare_for_class(_A , _A , return_labels=_A ) __lowerCAmelCase = model(**_A ).loss loss.backward() def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_A ), *get_values(_A ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): __lowerCAmelCase = problem_type["title"] __lowerCAmelCase = problem_type["num_labels"] __lowerCAmelCase = model_class(_A ) model.to(_A ) model.train() __lowerCAmelCase = self._prepare_for_class(_A , _A , return_labels=_A ) if problem_type["num_labels"] > 1: __lowerCAmelCase = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) __lowerCAmelCase = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_A ) as warning_list: __lowerCAmelCase = model(**_A ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = DeiTModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def _a ( ): __lowerCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a__ ( unittest.TestCase ): @cached_property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( _A ) __lowerCAmelCase = self.default_image_processor __lowerCAmelCase = prepare_img() __lowerCAmelCase = image_processor(images=_A , return_tensors="pt" ).to(_A ) # forward pass with torch.no_grad(): __lowerCAmelCase = model(**_A ) # verify the logits __lowerCAmelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _A ) __lowerCAmelCase = torch.tensor([-1.02_66, 0.19_12, -1.28_61] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) __lowerCAmelCase = self.default_image_processor __lowerCAmelCase = prepare_img() __lowerCAmelCase = image_processor(images=_A , return_tensors="pt" ) __lowerCAmelCase = inputs.pixel_values.to(_A ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __lowerCAmelCase = model(_A )
92
0
"""simple docstring""" def _snake_case ( UpperCAmelCase_ : int = 1000 ): A__ = 2**power A__ = 0 while n: A__ , A__ = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
69
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 SCREAMING_SNAKE_CASE_ : Any = data_utils.TransfoXLTokenizer SCREAMING_SNAKE_CASE_ : Union[str, Any] = data_utils.TransfoXLCorpus SCREAMING_SNAKE_CASE_ : str = data_utils SCREAMING_SNAKE_CASE_ : List[Any] = data_utils def _snake_case ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCAmelCase_ , """rb""" ) as fp: A__ = pickle.load(UpperCAmelCase_ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) A__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(F"""Save vocabulary to {pytorch_vocab_dump_path}""" ) A__ = corpus.vocab.__dict__ torch.save(UpperCAmelCase_ , UpperCAmelCase_ ) A__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCAmelCase_ ) A__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(F"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(UpperCAmelCase_ , UpperCAmelCase_ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model A__ = os.path.abspath(UpperCAmelCase_ ) A__ = os.path.abspath(UpperCAmelCase_ ) print(F"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": A__ = TransfoXLConfig() else: A__ = TransfoXLConfig.from_json_file(UpperCAmelCase_ ) print(F"""Building PyTorch model from configuration: {config}""" ) A__ = TransfoXLLMHeadModel(UpperCAmelCase_ ) A__ = load_tf_weights_in_transfo_xl(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # Save pytorch-model A__ = os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) A__ = os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) print(F"""Save PyTorch model to {os.path.abspath(UpperCAmelCase_ )}""" ) torch.save(model.state_dict() , UpperCAmelCase_ ) print(F"""Save configuration file to {os.path.abspath(UpperCAmelCase_ )}""" ) with open(UpperCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) SCREAMING_SNAKE_CASE_ : Any = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
69
1
from collections import namedtuple import requests from lxml import html # type: ignore SCREAMING_SNAKE_CASE :Union[str, Any] = namedtuple('''covid_data''', '''cases deaths recovered''') def _lowerCAmelCase ( lowerCAmelCase_ :str = "https://www.worldometers.info/coronavirus/" )->covid_data: '''simple docstring''' snake_case_ = "//div[@class = \"maincounter-number\"]/span/text()" return covid_data(*html.fromstring(requests.get(lowerCAmelCase_ ).content ).xpath(lowerCAmelCase_ ) ) SCREAMING_SNAKE_CASE :str = '''Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}''' print(fmt.format(*covid_stats()))
159
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def _lowerCAmelCase ( lowerCAmelCase_ :List[str] )->str: '''simple docstring''' snake_case_ = [] for line in lines: snake_case_ = re.sub(r"#.*" , "" , lowerCAmelCase_ ) # remove comments if line: filtered_lines.append(lowerCAmelCase_ ) snake_case_ = "\n".join(lowerCAmelCase_ ) # Make a hash from all this code snake_case_ = full_str.encode("utf-8" ) return shaaaa(lowerCAmelCase_ ).hexdigest() # get importable module names and hash for caching SCREAMING_SNAKE_CASE :Union[str, Any] = { '''csv''': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), '''json''': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), '''pandas''': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), '''parquet''': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), '''arrow''': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), '''text''': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), '''imagefolder''': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), '''audiofolder''': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions SCREAMING_SNAKE_CASE :str = { '''.csv''': ('''csv''', {}), '''.tsv''': ('''csv''', {'''sep''': '''\t'''}), '''.json''': ('''json''', {}), '''.jsonl''': ('''json''', {}), '''.parquet''': ('''parquet''', {}), '''.arrow''': ('''arrow''', {}), '''.txt''': ('''text''', {}), } _EXTENSION_TO_MODULE.update({ext: ('''imagefolder''', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('''imagefolder''', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('''audiofolder''', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('''audiofolder''', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) SCREAMING_SNAKE_CASE :List[Any] = {'''imagefolder''', '''audiofolder'''} # Used to filter data files based on extensions given a module name SCREAMING_SNAKE_CASE :Dict[str, List[str]] = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('''.zip''') _MODULE_TO_EXTENSIONS["audiofolder"].append('''.zip''')
159
1
import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : List[Any] ): __UpperCamelCase =VideoMAEConfig() set_architecture_configs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if "finetuned" not in model_name: __UpperCamelCase =False if "finetuned" in model_name: __UpperCamelCase ='huggingface/label-files' if "kinetics" in model_name: __UpperCamelCase =4_00 __UpperCamelCase ='kinetics400-id2label.json' elif "ssv2" in model_name: __UpperCamelCase =1_74 __UpperCamelCase ='something-something-v2-id2label.json' else: raise ValueError('Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.' ) __UpperCamelCase =json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type='dataset' ) , 'r' ) ) __UpperCamelCase ={int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} __UpperCamelCase =idalabel __UpperCamelCase ={v: k for k, v in idalabel.items()} return config def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[str] ): if "small" in model_name: __UpperCamelCase =3_84 __UpperCamelCase =15_36 __UpperCamelCase =12 __UpperCamelCase =16 __UpperCamelCase =12 __UpperCamelCase =3 __UpperCamelCase =1_92 __UpperCamelCase =7_68 elif "large" in model_name: __UpperCamelCase =10_24 __UpperCamelCase =40_96 __UpperCamelCase =24 __UpperCamelCase =16 __UpperCamelCase =12 __UpperCamelCase =8 __UpperCamelCase =5_12 __UpperCamelCase =20_48 elif "huge" in model_name: __UpperCamelCase =12_80 __UpperCamelCase =51_20 __UpperCamelCase =32 __UpperCamelCase =16 __UpperCamelCase =12 __UpperCamelCase =8 __UpperCamelCase =6_40 __UpperCamelCase =25_60 elif "base" not in model_name: raise ValueError('Model name should include either "small", "base", "large", or "huge"' ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Union[str, Any] ): if "encoder." in name: __UpperCamelCase =name.replace('encoder.' , '' ) if "cls_token" in name: __UpperCamelCase =name.replace('cls_token' , 'videomae.embeddings.cls_token' ) if "decoder_pos_embed" in name: __UpperCamelCase =name.replace('decoder_pos_embed' , 'decoder.decoder_pos_embed' ) if "pos_embed" in name and "decoder" not in name: __UpperCamelCase =name.replace('pos_embed' , 'videomae.embeddings.position_embeddings' ) if "patch_embed.proj" in name: __UpperCamelCase =name.replace('patch_embed.proj' , 'videomae.embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: __UpperCamelCase =name.replace('patch_embed.norm' , 'videomae.embeddings.norm' ) if "decoder.blocks" in name: __UpperCamelCase =name.replace('decoder.blocks' , 'decoder.decoder_layers' ) if "blocks" in name: __UpperCamelCase =name.replace('blocks' , 'videomae.encoder.layer' ) if "attn.proj" in name: __UpperCamelCase =name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name and "bias" not in name: __UpperCamelCase =name.replace('attn' , 'attention.self' ) if "attn" in name: __UpperCamelCase =name.replace('attn' , 'attention.attention' ) if "norm1" in name: __UpperCamelCase =name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: __UpperCamelCase =name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: __UpperCamelCase =name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: __UpperCamelCase =name.replace('mlp.fc2' , 'output.dense' ) if "decoder_embed" in name: __UpperCamelCase =name.replace('decoder_embed' , 'decoder.decoder_embed' ) if "decoder_norm" in name: __UpperCamelCase =name.replace('decoder_norm' , 'decoder.decoder_norm' ) if "decoder_pred" in name: __UpperCamelCase =name.replace('decoder_pred' , 'decoder.decoder_pred' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: __UpperCamelCase =name.replace('norm.weight' , 'videomae.layernorm.weight' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: __UpperCamelCase =name.replace('norm.bias' , 'videomae.layernorm.bias' ) if "head" in name and "decoder" not in name: __UpperCamelCase =name.replace('head' , 'classifier' ) return name def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict ): for key in orig_state_dict.copy().keys(): __UpperCamelCase =orig_state_dict.pop(SCREAMING_SNAKE_CASE__ ) if key.startswith('encoder.' ): __UpperCamelCase =key.replace('encoder.' , '' ) if "qkv" in key: __UpperCamelCase =key.split('.' ) if key.startswith('decoder.blocks' ): __UpperCamelCase =config.decoder_hidden_size __UpperCamelCase =int(key_split[2] ) __UpperCamelCase ='decoder.decoder_layers.' if "weight" in key: __UpperCamelCase =val[:dim, :] __UpperCamelCase =val[dim : dim * 2, :] __UpperCamelCase =val[-dim:, :] else: __UpperCamelCase =config.hidden_size __UpperCamelCase =int(key_split[1] ) __UpperCamelCase ='videomae.encoder.layer.' if "weight" in key: __UpperCamelCase =val[:dim, :] __UpperCamelCase =val[dim : dim * 2, :] __UpperCamelCase =val[-dim:, :] else: __UpperCamelCase =val return orig_state_dict def _UpperCAmelCase ( ): __UpperCamelCase =hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename='eating_spaghetti.npy' , repo_type='dataset' ) __UpperCamelCase =np.load(SCREAMING_SNAKE_CASE__ ) return list(SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple ): __UpperCamelCase =get_videomae_config(SCREAMING_SNAKE_CASE__ ) if "finetuned" in model_name: __UpperCamelCase =VideoMAEForVideoClassification(SCREAMING_SNAKE_CASE__ ) else: __UpperCamelCase =VideoMAEForPreTraining(SCREAMING_SNAKE_CASE__ ) # download original checkpoint, hosted on Google Drive __UpperCamelCase ='pytorch_model.bin' gdown.cached_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , quiet=SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =torch.load(SCREAMING_SNAKE_CASE__ , map_location='cpu' ) if "model" in files: __UpperCamelCase =files['model'] else: __UpperCamelCase =files['module'] __UpperCamelCase =convert_state_dict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) model.load_state_dict(SCREAMING_SNAKE_CASE__ ) model.eval() # verify model on basic input __UpperCamelCase =VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) __UpperCamelCase =prepare_video() __UpperCamelCase =image_processor(SCREAMING_SNAKE_CASE__ , return_tensors='pt' ) if "finetuned" not in model_name: __UpperCamelCase =hf_hub_download(repo_id='hf-internal-testing/bool-masked-pos' , filename='bool_masked_pos.pt' ) __UpperCamelCase =torch.load(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =model(**SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =outputs.logits __UpperCamelCase =[ 'videomae-small-finetuned-kinetics', 'videomae-small-finetuned-ssv2', # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) 'videomae-base-short', 'videomae-base-short-finetuned-kinetics', 'videomae-base', 'videomae-base-finetuned-kinetics', 'videomae-large', 'videomae-large-finetuned-kinetics', 'videomae-huge-finetuned-kinetics', # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) 'videomae-base-short-ssv2', 'videomae-base-short-finetuned-ssv2', 'videomae-base-ssv2', 'videomae-base-finetuned-ssv2', ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": __UpperCamelCase =torch.Size([1, 4_00] ) __UpperCamelCase =torch.tensor([-0.9291, -0.4061, -0.9307] ) elif model_name == "videomae-small-finetuned-ssv2": __UpperCamelCase =torch.Size([1, 1_74] ) __UpperCamelCase =torch.tensor([0.2671, -0.4689, -0.8235] ) elif model_name == "videomae-base": __UpperCamelCase =torch.Size([1, 14_08, 15_36] ) __UpperCamelCase =torch.tensor([[0.7739, 0.7968, 0.7089], [0.6701, 0.7487, 0.6209], [0.4287, 0.5158, 0.4773]] ) elif model_name == "videomae-base-short": __UpperCamelCase =torch.Size([1, 14_08, 15_36] ) __UpperCamelCase =torch.tensor([[0.7994, 0.9612, 0.8508], [0.7401, 0.8958, 0.8302], [0.5862, 0.7468, 0.7325]] ) # we verified the loss both for normalized and unnormalized targets for this one __UpperCamelCase =torch.tensor([0.5142] ) if config.norm_pix_loss else torch.tensor([0.6469] ) elif model_name == "videomae-large": __UpperCamelCase =torch.Size([1, 14_08, 15_36] ) __UpperCamelCase =torch.tensor([[0.7149, 0.7997, 0.6966], [0.6768, 0.7869, 0.6948], [0.5139, 0.6221, 0.5605]] ) elif model_name == "videomae-large-finetuned-kinetics": __UpperCamelCase =torch.Size([1, 4_00] ) __UpperCamelCase =torch.tensor([0.0771, 0.0011, -0.3625] ) elif model_name == "videomae-huge-finetuned-kinetics": __UpperCamelCase =torch.Size([1, 4_00] ) __UpperCamelCase =torch.tensor([0.2433, 0.1632, -0.4894] ) elif model_name == "videomae-base-short-finetuned-kinetics": __UpperCamelCase =torch.Size([1, 4_00] ) __UpperCamelCase =torch.tensor([0.6588, 0.0990, -0.2493] ) elif model_name == "videomae-base-finetuned-kinetics": __UpperCamelCase =torch.Size([1, 4_00] ) __UpperCamelCase =torch.tensor([0.3669, -0.0688, -0.2421] ) elif model_name == "videomae-base-short-ssv2": __UpperCamelCase =torch.Size([1, 14_08, 15_36] ) __UpperCamelCase =torch.tensor([[0.4712, 0.5296, 0.5786], [0.2278, 0.2729, 0.4026], [0.0352, 0.0730, 0.2506]] ) elif model_name == "videomae-base-short-finetuned-ssv2": __UpperCamelCase =torch.Size([1, 1_74] ) __UpperCamelCase =torch.tensor([-0.0537, -0.1539, -0.3266] ) elif model_name == "videomae-base-ssv2": __UpperCamelCase =torch.Size([1, 14_08, 15_36] ) __UpperCamelCase =torch.tensor([[0.8131, 0.8727, 0.8546], [0.7366, 0.9377, 0.8870], [0.5935, 0.8874, 0.8564]] ) elif model_name == "videomae-base-finetuned-ssv2": __UpperCamelCase =torch.Size([1, 1_74] ) __UpperCamelCase =torch.tensor([0.1961, -0.8337, -0.6389] ) else: raise ValueError(F'Model name not supported. Should be one of {model_names}' ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) else: print('Logits:' , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) print('Logits ok!' ) # verify loss, if applicable if model_name == "videomae-base-short": __UpperCamelCase =outputs.loss assert torch.allclose(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) print('Loss ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model and image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: print('Pushing to the hub...' ) model.push_to_hub(SCREAMING_SNAKE_CASE__ , organization='nielsr' ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) _A = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
368
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer _A = logging.get_logger(__name__) _A = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _A = { 'vocab_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/vocab.txt', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/vocab.txt', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt' ), 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt' ), 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt', 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json' ), 'bert-base-multilingual-cased': ( 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json' ), 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-cased': ( 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json' ), }, } _A = { 'bert-base-uncased': 512, 'bert-large-uncased': 512, 'bert-base-cased': 512, 'bert-large-cased': 512, 'bert-base-multilingual-uncased': 512, 'bert-base-multilingual-cased': 512, 'bert-base-chinese': 512, 'bert-base-german-cased': 512, 'bert-large-uncased-whole-word-masking': 512, 'bert-large-cased-whole-word-masking': 512, 'bert-large-uncased-whole-word-masking-finetuned-squad': 512, 'bert-large-cased-whole-word-masking-finetuned-squad': 512, 'bert-base-cased-finetuned-mrpc': 512, 'bert-base-german-dbmdz-cased': 512, 'bert-base-german-dbmdz-uncased': 512, 'TurkuNLP/bert-base-finnish-cased-v1': 512, 'TurkuNLP/bert-base-finnish-uncased-v1': 512, 'wietsedv/bert-base-dutch-cased': 512, } _A = { 'bert-base-uncased': {'do_lower_case': True}, 'bert-large-uncased': {'do_lower_case': True}, 'bert-base-cased': {'do_lower_case': False}, 'bert-large-cased': {'do_lower_case': False}, 'bert-base-multilingual-uncased': {'do_lower_case': True}, 'bert-base-multilingual-cased': {'do_lower_case': False}, 'bert-base-chinese': {'do_lower_case': False}, 'bert-base-german-cased': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking-finetuned-squad': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking-finetuned-squad': {'do_lower_case': False}, 'bert-base-cased-finetuned-mrpc': {'do_lower_case': False}, 'bert-base-german-dbmdz-cased': {'do_lower_case': False}, 'bert-base-german-dbmdz-uncased': {'do_lower_case': True}, 'TurkuNLP/bert-base-finnish-cased-v1': {'do_lower_case': False}, 'TurkuNLP/bert-base-finnish-uncased-v1': {'do_lower_case': True}, 'wietsedv/bert-base-dutch-cased': {'do_lower_case': False}, } class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Any = VOCAB_FILES_NAMES UpperCAmelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Optional[int] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Dict = BertTokenizer def __init__( self , A_=None , A_=None , A_=True , A_="[UNK]" , A_="[SEP]" , A_="[PAD]" , A_="[CLS]" , A_="[MASK]" , A_=True , A_=None , **A_ , ) -> Any: super().__init__( A_ , tokenizer_file=A_ , do_lower_case=A_ , unk_token=A_ , sep_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , tokenize_chinese_chars=A_ , strip_accents=A_ , **A_ , ) __UpperCamelCase =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , A_ ) != do_lower_case or normalizer_state.get('strip_accents' , A_ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , A_ ) != tokenize_chinese_chars ): __UpperCamelCase =getattr(A_ , normalizer_state.pop('type' ) ) __UpperCamelCase =do_lower_case __UpperCamelCase =strip_accents __UpperCamelCase =tokenize_chinese_chars __UpperCamelCase =normalizer_class(**A_ ) __UpperCamelCase =do_lower_case def _a ( self , A_ , A_=None ) -> List[str]: __UpperCamelCase =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _a ( self , A_ , A_ = None ) -> List[int]: __UpperCamelCase =[self.sep_token_id] __UpperCamelCase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _a ( self , A_ , A_ = None ) -> Tuple[str]: __UpperCamelCase =self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ )
117
0
"""simple docstring""" import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() def __a ( self ) -> Union[str, Any]: a, a : Tuple = FlaxControlNetModel.from_pretrained( "lllyasviel/sd-controlnet-canny" , from_pt=lowerCAmelCase__ , dtype=jnp.bfloataa ) a, a : int = FlaxStableDiffusionControlNetPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , controlnet=lowerCAmelCase__ , from_pt=lowerCAmelCase__ , dtype=jnp.bfloataa ) a : str = controlnet_params a : Union[str, Any] = "bird" a : List[str] = jax.device_count() a : Union[str, Any] = pipe.prepare_text_inputs([prompts] * num_samples ) a : Tuple = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png" ) a : str = pipe.prepare_image_inputs([canny_image] * num_samples ) a : Union[str, Any] = jax.random.PRNGKey(0 ) a : Any = jax.random.split(lowerCAmelCase__ , jax.device_count() ) a : Dict = replicate(lowerCAmelCase__ ) a : List[str] = shard(lowerCAmelCase__ ) a : List[Any] = shard(lowerCAmelCase__ ) a : Any = pipe( prompt_ids=lowerCAmelCase__ , image=lowerCAmelCase__ , params=lowerCAmelCase__ , prng_seed=lowerCAmelCase__ , num_inference_steps=50 , jit=lowerCAmelCase__ , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) a : str = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) a : Optional[int] = images[0, 253:256, 253:256, -1] a : str = jnp.asarray(jax.device_get(image_slice.flatten() ) ) a : Dict = jnp.array( [0.167_969, 0.116_699, 0.081_543, 0.154_297, 0.132_812, 0.108_887, 0.169_922, 0.169_922, 0.205_078] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def __a ( self ) -> str: a, a : Any = FlaxControlNetModel.from_pretrained( "lllyasviel/sd-controlnet-openpose" , from_pt=lowerCAmelCase__ , dtype=jnp.bfloataa ) a, a : Tuple = FlaxStableDiffusionControlNetPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , controlnet=lowerCAmelCase__ , from_pt=lowerCAmelCase__ , dtype=jnp.bfloataa ) a : Dict = controlnet_params a : Any = "Chef in the kitchen" a : Optional[Any] = jax.device_count() a : Union[str, Any] = pipe.prepare_text_inputs([prompts] * num_samples ) a : Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png" ) a : List[Any] = pipe.prepare_image_inputs([pose_image] * num_samples ) a : Tuple = jax.random.PRNGKey(0 ) a : Union[str, Any] = jax.random.split(lowerCAmelCase__ , jax.device_count() ) a : Dict = replicate(lowerCAmelCase__ ) a : List[str] = shard(lowerCAmelCase__ ) a : Any = shard(lowerCAmelCase__ ) a : Any = pipe( prompt_ids=lowerCAmelCase__ , image=lowerCAmelCase__ , params=lowerCAmelCase__ , prng_seed=lowerCAmelCase__ , num_inference_steps=50 , jit=lowerCAmelCase__ , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) a : Union[str, Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) a : Union[str, Any] = images[0, 253:256, 253:256, -1] a : Optional[Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) a : List[Any] = jnp.array( [[0.271_484, 0.261_719, 0.275_391, 0.277_344, 0.279_297, 0.291_016, 0.294_922, 0.302_734, 0.302_734]] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
105
"""simple docstring""" from __future__ import annotations from math import pi def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float , _lowercase : float ) ->dict[str, float]: '''simple docstring''' if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if inductance < 0: raise ValueError("Inductance cannot be negative" ) if frequency < 0: raise ValueError("Frequency cannot be negative" ) if reactance < 0: raise ValueError("Inductive reactance cannot be negative" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
105
1
'''simple docstring''' def __lowerCamelCase ( __snake_case : list ) -> list: """simple docstring""" if any(not isinstance(__snake_case, __snake_case ) or x < 0 for x in sequence ): raise TypeError("""Sequence must be list of non-negative integers""" ) for _ in range(len(__snake_case ) ): for i, (rod_upper, rod_lower) in enumerate(zip(__snake_case, sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
136
'''simple docstring''' import os import re import shutil import sys import tempfile import unittest import black __snake_case : List[str] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. __snake_case : Optional[int] = ' def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n' class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def lowercase__ ( self : Tuple ) -> Dict: '''simple docstring''' A__ : Tuple =tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , """models/bert/""" ) ) A__ : List[str] =self.transformer_dir shutil.copy( os.path.join(lowerCAmelCase_ , """src/transformers/models/bert/modeling_bert.py""" ) , os.path.join(self.transformer_dir , """models/bert/modeling_bert.py""" ) , ) def lowercase__ ( self : str ) -> List[str]: '''simple docstring''' A__ : Union[str, Any] ="""src/transformers""" shutil.rmtree(self.transformer_dir ) def lowercase__ ( self : Union[str, Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : int=None ) -> Dict: '''simple docstring''' A__ : Union[str, Any] =comment + f"\nclass {class_name}(nn.Module):\n" + class_code if overwrite_result is not None: A__ : Optional[Any] =comment + f"\nclass {class_name}(nn.Module):\n" + overwrite_result A__ : str =black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 ) A__ : Any =black.format_str(lowerCAmelCase_ , mode=lowerCAmelCase_ ) A__ : Union[str, Any] =os.path.join(self.transformer_dir , """new_code.py""" ) with open(lowerCAmelCase_ , """w""" , newline="""\n""" ) as f: f.write(lowerCAmelCase_ ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(lowerCAmelCase_ ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=lowerCAmelCase_ ) with open(lowerCAmelCase_ , """r""" ) as f: self.assertTrue(f.read() , lowerCAmelCase_ ) def lowercase__ ( self : Any ) -> Any: '''simple docstring''' A__ : Dict =check_copies.find_code_in_transformers("""models.bert.modeling_bert.BertLMPredictionHead""" ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase__ ( self : Dict ) -> List[Any]: '''simple docstring''' # Base copy consistency self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , REFERENCE_CODE + """\n""" , ) # With no empty line at the end self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , lowerCAmelCase_ , ) # Copy consistency with rename self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , re.sub("""Bert""" , """TestModel""" , lowerCAmelCase_ ) , ) # Copy consistency with a really long name A__ : Tuple ="""TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason""" self.check_copy_consistency( f"# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}" , f"{long_class_name}LMPredictionHead" , re.sub("""Bert""" , lowerCAmelCase_ , lowerCAmelCase_ ) , ) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , lowerCAmelCase_ , overwrite_result=re.sub("""Bert""" , """TestModel""" , lowerCAmelCase_ ) , ) def lowercase__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' A__ : List[Any] =check_copies.LOCALIZED_READMES["""README_zh-hans.md"""] A__ : List[str] =( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),""" """ released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**""" """ (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders""" """ as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang""" """ Luong, Quoc V. Le, Christopher D. Manning.""" ) A__ : Union[str, Any] =( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ : str =( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文""" """ [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自""" """ Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather""" """ than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,""" """ Christopher D. Manning 发布。\n""" ) A__ , A__ : Optional[Any] =check_copies.convert_to_localized_md( lowerCAmelCase_ , lowerCAmelCase_ , localized_readme["""format_model_list"""] ) self.assertFalse(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) A__ , A__ : List[str] =check_copies.convert_to_localized_md( lowerCAmelCase_ , lowerCAmelCase_ , localized_readme["""format_model_list"""] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(lowerCAmelCase_ ) A__ : Dict =( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.""" ) A__ : str =( """1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and""" """ the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ : int =( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ , A__ : int =check_copies.convert_to_localized_md( lowerCAmelCase_ , lowerCAmelCase_ , localized_readme["""format_model_list"""] ) # Check if the model link is synchronized. self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ )
136
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase=False ) -> Optional[int]: A: Any = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''deit.embeddings.cls_token'''), ('''dist_token''', '''deit.embeddings.distillation_token'''), ('''patch_embed.proj.weight''', '''deit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''deit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''deit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A: Dict = [(pair[0], pair[1][4:]) if pair[1].startswith('''deit''' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('''norm.weight''', '''deit.layernorm.weight'''), ('''norm.bias''', '''deit.layernorm.bias'''), ('''head.weight''', '''cls_classifier.weight'''), ('''head.bias''', '''cls_classifier.bias'''), ('''head_dist.weight''', '''distillation_classifier.weight'''), ('''head_dist.bias''', '''distillation_classifier.bias'''), ] ) return rename_keys def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase=False ) -> Optional[Any]: for i in range(config.num_hidden_layers ): if base_model: A: Tuple = '''''' else: A: Dict = '''deit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A: Tuple = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) A: Tuple = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A: Optional[int] = in_proj_weight[ : config.hidden_size, : ] A: Dict = in_proj_bias[: config.hidden_size] A: int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A: List[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A: int = in_proj_weight[ -config.hidden_size :, : ] A: Dict = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> Optional[int]: A: Any = dct.pop(__lowercase ) A: Any = val def SCREAMING_SNAKE_CASE( ) -> Optional[int]: A: List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' A: Optional[int] = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Any: A: Optional[Any] = DeiTConfig() # all deit models have fine-tuned heads A: int = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A: Union[str, Any] = 1_0_0_0 A: Union[str, Any] = '''huggingface/label-files''' A: List[str] = '''imagenet-1k-id2label.json''' A: Optional[int] = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type='''dataset''' ) , '''r''' ) ) A: List[Any] = {int(__lowercase ): v for k, v in idalabel.items()} A: Dict = idalabel A: Union[str, Any] = {v: k for k, v in idalabel.items()} A: Dict = int(deit_name[-6:-4] ) A: Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('''tiny''' ): A: Optional[int] = 1_9_2 A: Optional[Any] = 7_6_8 A: Optional[int] = 1_2 A: Union[str, Any] = 3 elif deit_name[9:].startswith('''small''' ): A: str = 3_8_4 A: List[str] = 1_5_3_6 A: int = 1_2 A: List[str] = 6 if deit_name[9:].startswith('''base''' ): pass elif deit_name[4:].startswith('''large''' ): A: Optional[Any] = 1_0_2_4 A: List[Any] = 4_0_9_6 A: Optional[int] = 2_4 A: Dict = 1_6 # load original model from timm A: int = timm.create_model(__lowercase , pretrained=__lowercase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A: int = timm_model.state_dict() A: List[str] = create_rename_keys(__lowercase , __lowercase ) for src, dest in rename_keys: rename_key(__lowercase , __lowercase , __lowercase ) read_in_q_k_v(__lowercase , __lowercase , __lowercase ) # load HuggingFace model A: Any = DeiTForImageClassificationWithTeacher(__lowercase ).eval() model.load_state_dict(__lowercase ) # Check outputs on an image, prepared by DeiTImageProcessor A: Optional[int] = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A: List[str] = DeiTImageProcessor(size=__lowercase , crop_size=config.image_size ) A: Optional[int] = image_processor(images=prepare_img() , return_tensors='''pt''' ) A: Any = encoding['''pixel_values'''] A: int = model(__lowercase ) A: str = timm_model(__lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowercase , outputs.logits , atol=1E-3 ) Path(__lowercase ).mkdir(exist_ok=__lowercase ) print(F"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowercase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowercase ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) UpperCamelCase = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
319
'''simple docstring''' import heapq import sys import numpy as np UpperCamelCase = tuple[int, int] class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[Any] ) -> str: '''simple docstring''' A: Any = [] A: int = set() def _snake_case ( self : Optional[Any] ) -> int: '''simple docstring''' if not self.empty(): return self.elements[0][0] else: return float('''inf''' ) def _snake_case ( self : List[str] ) -> List[Any]: '''simple docstring''' return len(self.elements ) == 0 def _snake_case ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Any ) -> List[Any]: '''simple docstring''' if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(SCREAMING_SNAKE_CASE_ ) else: # update # print("update", item) A: Optional[int] = [] ((A) , (A)): str = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((A) , (A)): int = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def _snake_case ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : str ) -> Any: '''simple docstring''' if item in self.set: self.set.remove(SCREAMING_SNAKE_CASE_ ) A: str = [] ((A) , (A)): List[str] = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((A) , (A)): Any = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def _snake_case ( self : List[Any] ) -> Optional[int]: '''simple docstring''' return self.elements[0][1] def _snake_case ( self : int ) -> Union[str, Any]: '''simple docstring''' ((A) , (A)): Dict = heapq.heappop(self.elements ) self.set.remove(SCREAMING_SNAKE_CASE_ ) return (priority, item) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Union[str, Any]: # euclidean distance A: List[str] = np.array(__lowercase ) A: Optional[int] = np.array(__lowercase ) return np.linalg.norm(a - b ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> int: # integer division by time variable return consistent_heuristic(__lowercase , __lowercase ) // t def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Optional[Any]: # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase ) -> List[Any]: A: int = g_function[start] + Wa * heuristics[i](__lowercase , __lowercase ) return ans def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> Optional[int]: A: Union[str, Any] = np.chararray((n, n) ) for i in range(__lowercase ): for j in range(__lowercase ): A: Union[str, Any] = '''*''' for i in range(__lowercase ): for j in range(__lowercase ): if (j, (n - 1) - i) in blocks: A: Optional[Any] = '''#''' A: Tuple = '''-''' A: List[str] = back_pointer[goal] while x != start: ((A) , (A)): Tuple = x # print(x) A: List[str] = '''-''' A: str = back_pointer[x] A: Dict = '''-''' for i in range(__lowercase ): for j in range(__lowercase ): if (i, j) == (0, n - 1): print(grid[i][j] , end=''' ''' ) print('''<-- End position''' , end=''' ''' ) else: print(grid[i][j] , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) print('''PATH TAKEN BY THE ALGORITHM IS:-''' ) A: List[str] = back_pointer[goal] while x != start: print(__lowercase , end=''' ''' ) A: Optional[int] = back_pointer[x] print(__lowercase ) sys.exit() def SCREAMING_SNAKE_CASE( __lowercase ) -> Optional[Any]: if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Union[str, Any]: for itera in range(__lowercase ): open_list[itera].remove_element(__lowercase ) # print("s", s) # print("j", j) ((A) , (A)): Tuple = s A: Optional[Any] = (x - 1, y) A: str = (x + 1, y) A: List[Any] = (x, y + 1) A: int = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(__lowercase ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(__lowercase ) A: int = -1 A: int = float('''inf''' ) if valid(__lowercase ) and g_function[neighbours] > g_function[s] + 1: A: List[str] = g_function[s] + 1 A: List[str] = s if neighbours not in close_list_anchor: open_list[0].put(__lowercase , key(__lowercase , 0 , __lowercase , __lowercase ) ) if neighbours not in close_list_inad: for var in range(1 , __lowercase ): if key(__lowercase , __lowercase , __lowercase , __lowercase ) <= Wa * key( __lowercase , 0 , __lowercase , __lowercase ): open_list[j].put( __lowercase , key(__lowercase , __lowercase , __lowercase , __lowercase ) ) def SCREAMING_SNAKE_CASE( ) -> Tuple: A: str = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list UpperCamelCase = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} UpperCamelCase = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] UpperCamelCase = make_common_ground() UpperCamelCase = blocks_blk # hyper parameters UpperCamelCase = 1 UpperCamelCase = 1 UpperCamelCase = 20 UpperCamelCase = 3 # one consistent and two other inconsistent # start and end destination UpperCamelCase = (0, 0) UpperCamelCase = (n - 1, n - 1) UpperCamelCase = 1 def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> int: A: int = {start: 0, goal: float('''inf''' )} A: Union[str, Any] = {start: -1, goal: -1} A: List[Any] = [] A: Union[str, Any] = set() for i in range(__lowercase ): open_list.append(PriorityQueue() ) open_list[i].put(__lowercase , key(__lowercase , __lowercase , __lowercase , __lowercase ) ) A: list[int] = [] A: list[int] = [] while open_list[0].minkey() < float('''inf''' ): for i in range(1 , __lowercase ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float('''inf''' ): do_something(__lowercase , __lowercase , __lowercase ) else: A , A: Union[str, Any] = open_list[i].top_show() visited.add(__lowercase ) expand_state( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) close_list_inad.append(__lowercase ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float('''inf''' ): do_something(__lowercase , __lowercase , __lowercase ) else: A: Union[str, Any] = open_list[0].top_show() visited.add(__lowercase ) expand_state( __lowercase , 0 , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) close_list_anchor.append(__lowercase ) print('''No path found to goal''' ) print() for i in range(n - 1 , -1 , -1 ): for j in range(__lowercase ): if (j, i) in blocks: print('''#''' , end=''' ''' ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print('''*''' , end=''' ''' ) else: print('''-''' , end=''' ''' ) else: print('''*''' , end=''' ''' ) if (j, i) == (n - 1, n - 1): print('''<-- End position''' , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
319
1
"""simple docstring""" from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = analyze_text(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = list(" " + ascii_lowercase ) # what is our total sum of probabilities. __SCREAMING_SNAKE_CASE = sum(single_char_strings.values() ) # one length string __SCREAMING_SNAKE_CASE = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: __SCREAMING_SNAKE_CASE = single_char_strings[ch] __SCREAMING_SNAKE_CASE = my_str / all_sum my_fir_sum += prob * math.loga(lowerCAmelCase_ ) # entropy formula. # print entropy print(f"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string __SCREAMING_SNAKE_CASE = sum(two_char_strings.values() ) __SCREAMING_SNAKE_CASE = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: __SCREAMING_SNAKE_CASE = cha + cha if sequence in two_char_strings: __SCREAMING_SNAKE_CASE = two_char_strings[sequence] __SCREAMING_SNAKE_CASE = int(lowerCAmelCase_ ) / all_sum my_sec_sum += prob * math.loga(lowerCAmelCase_ ) # print second entropy print(f"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(f"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = Counter() # type: ignore __SCREAMING_SNAKE_CASE = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(lowerCAmelCase_ ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def UpperCAmelCase__ (): '''simple docstring''' import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
195
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' if digit_amount > 0: return round(number - int(lowerCAmelCase_ ) , lowerCAmelCase_ ) return number - int(lowerCAmelCase_ ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.3_45, 1)) print(decimal_isolate(35.3_45, 2)) print(decimal_isolate(35.3_45, 3)) print(decimal_isolate(-14.7_89, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.1_23, 1)) print(decimal_isolate(-14.1_23, 2)) print(decimal_isolate(-14.1_23, 3))
195
1
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _lowerCAmelCase : Optional[int] = logging.get_logger(__name__) _lowerCAmelCase : Union[str, Any] = { '''Salesforce/codegen-350M-nl''': '''https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json''', '''Salesforce/codegen-350M-multi''': '''https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json''', '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json''', '''Salesforce/codegen-2B-nl''': '''https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json''', '''Salesforce/codegen-2B-multi''': '''https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json''', '''Salesforce/codegen-2B-mono''': '''https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json''', '''Salesforce/codegen-6B-nl''': '''https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json''', '''Salesforce/codegen-6B-multi''': '''https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json''', '''Salesforce/codegen-6B-mono''': '''https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json''', '''Salesforce/codegen-16B-nl''': '''https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json''', '''Salesforce/codegen-16B-multi''': '''https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json''', '''Salesforce/codegen-16B-mono''': '''https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json''', } class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = '''codegen''' __UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self :Optional[Any] , snake_case :Tuple=50_400 , snake_case :Any=2_048 , snake_case :Tuple=2_048 , snake_case :str=4_096 , snake_case :List[Any]=28 , snake_case :str=16 , snake_case :Union[str, Any]=64 , snake_case :Tuple=None , snake_case :Optional[int]="gelu_new" , snake_case :List[Any]=0.0 , snake_case :Any=0.0 , snake_case :List[Any]=0.0 , snake_case :int=1e-5 , snake_case :List[str]=0.02 , snake_case :Dict=True , snake_case :Any=50_256 , snake_case :Any=50_256 , snake_case :Any=False , **snake_case :List[str] , ): '''simple docstring''' A_ : Any = vocab_size A_ : List[Any] = n_ctx A_ : Union[str, Any] = n_positions A_ : Optional[Any] = n_embd A_ : Optional[Any] = n_layer A_ : str = n_head A_ : int = n_inner A_ : int = rotary_dim A_ : List[str] = activation_function A_ : Optional[Any] = resid_pdrop A_ : List[str] = embd_pdrop A_ : int = attn_pdrop A_ : Any = layer_norm_epsilon A_ : Dict = initializer_range A_ : int = use_cache A_ : int = bos_token_id A_ : List[str] = eos_token_id super().__init__( bos_token_id=snake_case , eos_token_id=snake_case , tie_word_embeddings=snake_case , **snake_case ) class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" def __init__( self :Any , snake_case :PretrainedConfig , snake_case :str = "default" , snake_case :List[PatchingSpec] = None , snake_case :bool = False , ): '''simple docstring''' super().__init__(snake_case , task=snake_case , patching_specs=snake_case , use_past=snake_case ) if not getattr(self._config , "pad_token_id" , snake_case ): # TODO: how to do that better? A_ : Tuple = 0 @property def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : Dict = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(snake_case , direction="inputs" ) A_ : List[str] = {0: "batch", 1: "past_sequence + sequence"} else: A_ : Optional[int] = {0: "batch", 1: "sequence"} return common_inputs @property def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' return self._config.n_layer @property def SCREAMING_SNAKE_CASE ( self :Union[str, Any] ): '''simple docstring''' return self._config.n_head def SCREAMING_SNAKE_CASE ( self :int , snake_case :PreTrainedTokenizer , snake_case :int = -1 , snake_case :int = -1 , snake_case :bool = False , snake_case :Optional[TensorType] = None , ): '''simple docstring''' A_ : Any = super(snake_case , self ).generate_dummy_inputs( snake_case , batch_size=snake_case , seq_length=snake_case , is_pair=snake_case , framework=snake_case ) # We need to order the input in the way they appears in the forward() A_ : int = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A_ , A_ : int = common_inputs["input_ids"].shape # Not using the same length for past_key_values A_ : Optional[int] = seqlen + 2 A_ : Any = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) A_ : Dict = [ (torch.zeros(snake_case ), torch.zeros(snake_case )) for _ in range(self.num_layers ) ] A_ : Optional[int] = common_inputs["attention_mask"] if self.use_past: A_ : int = ordered_inputs["attention_mask"].dtype A_ : Any = torch.cat( [ordered_inputs["attention_mask"], torch.ones(snake_case , snake_case , dtype=snake_case )] , dim=1 ) return ordered_inputs @property def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' return 13
300
def __snake_case ( _lowerCAmelCase : List[str] , _lowerCAmelCase : int ) -> str: return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any]=0 ) -> Any: return sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : x[column] ) def __snake_case ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : Any=float("inf" ) ) -> int: for i in range(points_counts - 1 ): for j in range(i + 1 , _lowerCAmelCase ): A_ : Tuple = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: A_ : Union[str, Any] = current_dis return min_dis def __snake_case ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str]=float("inf" ) ) -> Dict: for i in range(min(6 , points_counts - 1 ) , _lowerCAmelCase ): for j in range(max(0 , i - 6 ) , _lowerCAmelCase ): A_ : List[Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: A_ : Union[str, Any] = current_dis return min_dis def __snake_case ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Dict ) -> List[str]: # base case if points_counts <= 3: return dis_between_closest_pair(_lowerCAmelCase , _lowerCAmelCase ) # recursion A_ : Optional[int] = points_counts // 2 A_ : List[Any] = closest_pair_of_points_sqr( _lowerCAmelCase , points_sorted_on_y[:mid] , _lowerCAmelCase ) A_ : List[Any] = closest_pair_of_points_sqr( _lowerCAmelCase , points_sorted_on_y[mid:] , points_counts - mid ) A_ : Tuple = min(_lowerCAmelCase , _lowerCAmelCase ) A_ : Dict = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(_lowerCAmelCase ) A_ : Tuple = dis_between_closest_in_strip( _lowerCAmelCase , len(_lowerCAmelCase ) , _lowerCAmelCase ) return min(_lowerCAmelCase , _lowerCAmelCase ) def __snake_case ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] ) -> Any: A_ : Optional[Any] = column_based_sort(_lowerCAmelCase , column=0 ) A_ : Optional[int] = column_based_sort(_lowerCAmelCase , column=1 ) return ( closest_pair_of_points_sqr( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) ) ** 0.5 if __name__ == "__main__": _lowerCAmelCase : List[Any] = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print('''Distance:''', closest_pair_of_points(points, len(points)))
300
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
1
from collections.abc import Sequence from queue import Queue class A : '''simple docstring''' def __init__( self : Optional[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Dict , __lowerCAmelCase : Union[str, Any]=None , __lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" A__ = start A__ = end A__ = val A__ = (start + end) // 2 A__ = left A__ = right def __repr__( self : int ) -> List[str]: """simple docstring""" return f'SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})' class A : '''simple docstring''' def __init__( self : Optional[int] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[Any] ) -> str: """simple docstring""" A__ = collection A__ = function if self.collection: A__ = self._build_tree(0 , len(__snake_case ) - 1 ) def a_ ( self : Dict , __lowerCAmelCase : str , __lowerCAmelCase : List[Any] ) -> Dict: """simple docstring""" self._update_tree(self.root , __snake_case , __snake_case ) def a_ ( self : int , __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[Any] ) -> str: """simple docstring""" return self._query_range(self.root , __snake_case , __snake_case ) def a_ ( self : str , __lowerCAmelCase : Dict , __lowerCAmelCase : int ) -> str: """simple docstring""" if start == end: return SegmentTreeNode(__snake_case , __snake_case , self.collection[start] ) A__ = (start + end) // 2 A__ = self._build_tree(__snake_case , __snake_case ) A__ = self._build_tree(mid + 1 , __snake_case ) return SegmentTreeNode(__snake_case , __snake_case , self.fn(left.val , right.val ) , __snake_case , __snake_case ) def a_ ( self : Tuple , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Union[str, Any] ) -> Tuple: """simple docstring""" if node.start == i and node.end == i: A__ = val return if i <= node.mid: self._update_tree(node.left , __snake_case , __snake_case ) else: self._update_tree(node.right , __snake_case , __snake_case ) A__ = self.fn(node.left.val , node.right.val ) def a_ ( self : Optional[int] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Union[str, Any] ) -> int: """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , __snake_case , __snake_case ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , __snake_case , node.mid ) , self._query_range(node.right , node.mid + 1 , __snake_case ) , ) else: # range in right child tree return self._query_range(node.right , __snake_case , __snake_case ) def a_ ( self : List[Any] ) -> List[str]: """simple docstring""" if self.root is not None: A__ = Queue() queue.put(self.root ) while not queue.empty(): A__ = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 5_0) A : Dict = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
274
class A__ : """simple docstring""" def __init__( self , __snake_case , __snake_case , __snake_case ): snake_case = name snake_case = value snake_case = weight def __repr__( self ): return F'''{self.__class__.__name__}({self.name}, {self.value}, {self.weight})''' def a_ ( self ): return self.value def a_ ( self ): return self.name def a_ ( self ): return self.weight def a_ ( self ): return self.value / self.weight def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" snake_case = [] for i in range(len(UpperCamelCase_ ) ): menu.append(Things(name[i] ,value[i] ,weight[i] ) ) return menu def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" snake_case = sorted(UpperCamelCase_ ,key=UpperCamelCase_ ,reverse=UpperCamelCase_ ) snake_case = [] snake_case , snake_case = 0.0, 0.0 for i in range(len(UpperCamelCase_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCAmelCase__ (): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
127
0
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: _A : List[str] =None _A : Tuple =logging.get_logger(__name__) _A : List[Any] ={'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} _A : Optional[Any] ={ '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/mbart-large-en-ro''': '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json''', '''facebook/mbart-large-cc25''': '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json''', }, } _A : Any ={ '''facebook/mbart-large-en-ro''': 1_024, '''facebook/mbart-large-cc25''': 1_024, } # fmt: off _A : Optional[int] =['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class _lowercase ( _lowercase ): a = VOCAB_FILES_NAMES a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a = PRETRAINED_VOCAB_FILES_MAP a = ["""input_ids""", """attention_mask"""] a = MBartTokenizer a = [] a = [] def __init__( self: int , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: Tuple=None , UpperCamelCase__: int="<s>" , UpperCamelCase__: Dict="</s>" , UpperCamelCase__: Tuple="</s>" , UpperCamelCase__: Union[str, Any]="<s>" , UpperCamelCase__: Any="<unk>" , UpperCamelCase__: List[Any]="<pad>" , UpperCamelCase__: Union[str, Any]="<mask>" , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: Any=None , UpperCamelCase__: List[str]=None , **UpperCamelCase__: Optional[int] , ): # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( vocab_file=UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , src_lang=UpperCamelCase__ , tgt_lang=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase__ : Dict = vocab_file lowerCamelCase__ : Optional[Any] = False if not self.vocab_file else True lowerCamelCase__ : List[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) lowerCamelCase__ : Dict = { lang_code: self.convert_tokens_to_ids(UpperCamelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : Optional[int] = src_lang if src_lang is not None else """en_XX""" lowerCamelCase__ : List[str] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowerCamelCase_ ( self: Optional[Any] ): return self._src_lang @src_lang.setter def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): lowerCamelCase__ : Tuple = [self.sep_token_id] lowerCamelCase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: str , UpperCamelCase__: Optional[str] , UpperCamelCase__: Optional[str] , **UpperCamelCase__: List[Any] ): if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) lowerCamelCase__ : Any = src_lang lowerCamelCase__ : List[str] = self(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Dict = self.convert_tokens_to_ids(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tgt_lang_id return inputs def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: str = "en_XX" , UpperCamelCase__: Optional[List[str]] = None , UpperCamelCase__: str = "ro_RO" , **UpperCamelCase__: Optional[Any] , ): lowerCamelCase__ : Tuple = src_lang lowerCamelCase__ : Any = tgt_lang return super().prepare_seqaseq_batch(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): return self.set_src_lang_special_tokens(self.src_lang ) def lowerCamelCase_ ( self: List[str] ): return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Any ): lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(UpperCamelCase__ ) lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] lowerCamelCase__ : Optional[int] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : int = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : Optional[int] = self.convert_tokens_to_ids(UpperCamelCase__ ) lowerCamelCase__ : Tuple = [] lowerCamelCase__ : int = [self.eos_token_id, self.cur_lang_code] lowerCamelCase__ : List[str] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[int] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowerCamelCase_ ( self: int , UpperCamelCase__: str , UpperCamelCase__: Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory.''' ) return lowerCamelCase__ : int = os.path.join( UpperCamelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
363
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[Any]: if "cls_token" in name: lowerCamelCase__ : Any = name.replace("""cls_token""" , """vit.embeddings.cls_token""" ) if "mask_token" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""mask_token""" , """decoder.mask_token""" ) if "decoder_pos_embed" in name: lowerCamelCase__ : Tuple = name.replace("""decoder_pos_embed""" , """decoder.decoder_pos_embed""" ) if "pos_embed" in name and "decoder" not in name: lowerCamelCase__ : str = name.replace("""pos_embed""" , """vit.embeddings.position_embeddings""" ) if "patch_embed.proj" in name: lowerCamelCase__ : Optional[int] = name.replace("""patch_embed.proj""" , """vit.embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: lowerCamelCase__ : Any = name.replace("""patch_embed.norm""" , """vit.embeddings.norm""" ) if "decoder_blocks" in name: lowerCamelCase__ : Dict = name.replace("""decoder_blocks""" , """decoder.decoder_layers""" ) if "blocks" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""blocks""" , """vit.encoder.layer""" ) if "attn.proj" in name: lowerCamelCase__ : List[Any] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: lowerCamelCase__ : List[str] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: lowerCamelCase__ : Any = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: lowerCamelCase__ : str = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""mlp.fc2""" , """output.dense""" ) if "decoder_embed" in name: lowerCamelCase__ : Tuple = name.replace("""decoder_embed""" , """decoder.decoder_embed""" ) if "decoder_norm" in name: lowerCamelCase__ : Optional[int] = name.replace("""decoder_norm""" , """decoder.decoder_norm""" ) if "decoder_pred" in name: lowerCamelCase__ : int = name.replace("""decoder_pred""" , """decoder.decoder_pred""" ) if "norm.weight" in name and "decoder" not in name: lowerCamelCase__ : Union[str, Any] = name.replace("""norm.weight""" , """vit.layernorm.weight""" ) if "norm.bias" in name and "decoder" not in name: lowerCamelCase__ : Dict = name.replace("""norm.bias""" , """vit.layernorm.bias""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : List[str] = orig_state_dict.pop(UpperCamelCase ) if "qkv" in key: lowerCamelCase__ : List[Any] = key.split(""".""" ) lowerCamelCase__ : Optional[int] = int(key_split[1] ) if "decoder_blocks" in key: lowerCamelCase__ : str = config.decoder_hidden_size lowerCamelCase__ : List[Any] = """decoder.decoder_layers.""" if "weight" in key: lowerCamelCase__ : int = val[:dim, :] lowerCamelCase__ : int = val[dim : dim * 2, :] lowerCamelCase__ : Tuple = val[-dim:, :] elif "bias" in key: lowerCamelCase__ : Tuple = val[:dim] lowerCamelCase__ : Optional[int] = val[dim : dim * 2] lowerCamelCase__ : List[Any] = val[-dim:] else: lowerCamelCase__ : List[Any] = config.hidden_size lowerCamelCase__ : Optional[int] = """vit.encoder.layer.""" if "weight" in key: lowerCamelCase__ : str = val[:dim, :] lowerCamelCase__ : List[Any] = val[dim : dim * 2, :] lowerCamelCase__ : Optional[int] = val[-dim:, :] elif "bias" in key: lowerCamelCase__ : int = val[:dim] lowerCamelCase__ : List[Any] = val[dim : dim * 2] lowerCamelCase__ : Optional[int] = val[-dim:] else: lowerCamelCase__ : int = val return orig_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Any = ViTMAEConfig() if "large" in checkpoint_url: lowerCamelCase__ : Any = 1024 lowerCamelCase__ : Optional[Any] = 4096 lowerCamelCase__ : List[str] = 24 lowerCamelCase__ : Union[str, Any] = 16 elif "huge" in checkpoint_url: lowerCamelCase__ : List[str] = 14 lowerCamelCase__ : Dict = 1280 lowerCamelCase__ : Tuple = 5120 lowerCamelCase__ : List[str] = 32 lowerCamelCase__ : Union[str, Any] = 16 lowerCamelCase__ : List[Any] = ViTMAEForPreTraining(UpperCamelCase ) lowerCamelCase__ : str = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""model"""] lowerCamelCase__ : Union[str, Any] = ViTMAEImageProcessor(size=config.image_size ) lowerCamelCase__ : List[str] = convert_state_dict(UpperCamelCase , UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() lowerCamelCase__ : Union[str, Any] = """https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg""" lowerCamelCase__ : List[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) lowerCamelCase__ : str = ViTMAEImageProcessor(size=config.image_size ) lowerCamelCase__ : Any = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) lowerCamelCase__ : Optional[Any] = model(**UpperCamelCase ) lowerCamelCase__ : Optional[Any] = outputs.logits if "large" in checkpoint_url: lowerCamelCase__ : List[Any] = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: lowerCamelCase__ : Optional[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: lowerCamelCase__ : int = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , UpperCamelCase , atol=1E-4 ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
129
0
'''simple docstring''' from __future__ import annotations import collections import pprint from pathlib import Path def _UpperCamelCase ( UpperCamelCase__ ): return "".join(sorted(UpperCamelCase__ ) ) def _UpperCamelCase ( UpperCamelCase__ ): return word_by_signature[signature(UpperCamelCase__ )] __A =Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') __A =sorted({word.strip().lower() for word in data.splitlines()}) __A =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": __A ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
163
'''simple docstring''' import os import sys import unittest _lowerCamelCase : Optional[int] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) _lowerCamelCase : str = os.path.join('tests', 'models', 'bert', 'test_modeling_bert.py') _lowerCamelCase : Tuple = os.path.join('tests', 'models', 'blip', 'test_modeling_blip.py') class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def A (self : Any ): A = get_test_to_tester_mapping(_lowerCAmelCase ) A = get_test_to_tester_mapping(_lowerCAmelCase ) A = {"""BertModelTest""": """BertModelTester"""} A = { """BlipModelTest""": """BlipModelTester""", """BlipTextImageModelTest""": """BlipTextImageModelsModelTester""", """BlipTextModelTest""": """BlipTextModelTester""", """BlipTextRetrievalModelTest""": """BlipTextRetrievalModelTester""", """BlipVQAModelTest""": """BlipVQAModelTester""", """BlipVisionModelTest""": """BlipVisionModelTester""", } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def A (self : str ): A = get_model_to_test_mapping(_lowerCAmelCase ) A = get_model_to_test_mapping(_lowerCAmelCase ) A = { """BertForMaskedLM""": ["""BertModelTest"""], """BertForMultipleChoice""": ["""BertModelTest"""], """BertForNextSentencePrediction""": ["""BertModelTest"""], """BertForPreTraining""": ["""BertModelTest"""], """BertForQuestionAnswering""": ["""BertModelTest"""], """BertForSequenceClassification""": ["""BertModelTest"""], """BertForTokenClassification""": ["""BertModelTest"""], """BertLMHeadModel""": ["""BertModelTest"""], """BertModel""": ["""BertModelTest"""], } A = { """BlipForConditionalGeneration""": ["""BlipTextImageModelTest"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTest"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTest"""], """BlipModel""": ["""BlipModelTest"""], """BlipTextModel""": ["""BlipTextModelTest"""], """BlipVisionModel""": ["""BlipVisionModelTest"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def A (self : Union[str, Any] ): A = get_model_to_tester_mapping(_lowerCAmelCase ) A = get_model_to_tester_mapping(_lowerCAmelCase ) A = { """BertForMaskedLM""": ["""BertModelTester"""], """BertForMultipleChoice""": ["""BertModelTester"""], """BertForNextSentencePrediction""": ["""BertModelTester"""], """BertForPreTraining""": ["""BertModelTester"""], """BertForQuestionAnswering""": ["""BertModelTester"""], """BertForSequenceClassification""": ["""BertModelTester"""], """BertForTokenClassification""": ["""BertModelTester"""], """BertLMHeadModel""": ["""BertModelTester"""], """BertModel""": ["""BertModelTester"""], } A = { """BlipForConditionalGeneration""": ["""BlipTextImageModelsModelTester"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTester"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTester"""], """BlipModel""": ["""BlipModelTester"""], """BlipTextModel""": ["""BlipTextModelTester"""], """BlipVisionModel""": ["""BlipVisionModelTester"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase )
258
0
'''simple docstring''' import argparse import os import re import packaging.version a__ : List[Any] = 'examples/' a__ : Union[str, Any] = { 'examples': (re.compile(R'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(R'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(R'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), R'\1version="VERSION",'), 'doc': (re.compile(R'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } a__ : Any = { 'init': 'src/transformers/__init__.py', 'setup': 'setup.py', } a__ : List[str] = 'README.md' def _lowercase ( __A ,__A ,__A ): '''simple docstring''' with open(__A ,"""r""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: __UpperCamelCase = f.read() __UpperCamelCase , __UpperCamelCase = REPLACE_PATTERNS[pattern] __UpperCamelCase = replace.replace("""VERSION""" ,__A ) __UpperCamelCase = re_pattern.sub(__A ,__A ) with open(__A ,"""w""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: f.write(__A ) def _lowercase ( __A ): '''simple docstring''' for folder, directories, fnames in os.walk(__A ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__A ,__A ) ,__A ,pattern="""examples""" ) def _lowercase ( __A ,__A=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__A ,__A ,__A ) if not patch: update_version_in_examples(__A ) def _lowercase ( ): '''simple docstring''' __UpperCamelCase = """🤗 Transformers currently provides the following architectures""" __UpperCamelCase = """1. Want to contribute a new model?""" with open(__A ,"""r""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: __UpperCamelCase = f.readlines() # Find the start of the list. __UpperCamelCase = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 __UpperCamelCase = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): __UpperCamelCase = lines[index].replace( """https://huggingface.co/docs/transformers/main/model_doc""" ,"""https://huggingface.co/docs/transformers/model_doc""" ,) index += 1 with open(__A ,"""w""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: f.writelines(__A ) def _lowercase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] ,"""r""" ) as f: __UpperCamelCase = f.read() __UpperCamelCase = REPLACE_PATTERNS["""init"""][0].search(__A ).groups()[0] return packaging.version.parse(__A ) def _lowercase ( __A=False ): '''simple docstring''' __UpperCamelCase = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: __UpperCamelCase = default_version.base_version elif patch: __UpperCamelCase = f"{default_version.major}.{default_version.minor}.{default_version.micro + 1}" else: __UpperCamelCase = f"{default_version.major}.{default_version.minor + 1}.0" # Now let's ask nicely if that's the right one. __UpperCamelCase = input(f"Which version are you releasing? [{default_version}]" ) if len(__A ) == 0: __UpperCamelCase = default_version print(f"Updating version to {version}." ) global_version_update(__A ,patch=__A ) if not patch: print("""Cleaning main README, don't forget to run `make fix-copies`.""" ) clean_main_ref_in_model_list() def _lowercase ( ): '''simple docstring''' __UpperCamelCase = get_version() __UpperCamelCase = f"{current_version.major}.{current_version.minor + 1}.0.dev0" __UpperCamelCase = current_version.base_version # Check with the user we got that right. __UpperCamelCase = input(f"Which version are we developing now? [{dev_version}]" ) if len(__A ) == 0: __UpperCamelCase = dev_version print(f"Updating version to {version}." ) global_version_update(__A ) print("""Cleaning main README, don't forget to run `make fix-copies`.""" ) clean_main_ref_in_model_list() if __name__ == "__main__": a__ : Dict = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') a__ : Any = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
243
'''simple docstring''' import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class UpperCAmelCase__ ( UpperCAmelCase_): def __init__( self , lowercase=0.01 , lowercase=1_0_0_0 ) -> List[Any]: __UpperCamelCase = p_stop __UpperCamelCase = max_length def __iter__( self ) -> Dict: __UpperCamelCase = 0 __UpperCamelCase = False while not stop and count < self.max_length: yield count count += 1 __UpperCamelCase = random.random() < self.p_stop class UpperCAmelCase__ ( unittest.TestCase): def __lowerCamelCase ( self , lowercase , lowercase , lowercase=False , lowercase=True ) -> List[str]: __UpperCamelCase = [ BatchSamplerShard(lowercase , 2 , lowercase , split_batches=lowercase , even_batches=lowercase ) for i in range(2 ) ] __UpperCamelCase = [list(lowercase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowercase ) for shard in batch_sampler_shards] , [len(lowercase ) for e in expected] ) self.assertListEqual(lowercase , lowercase ) def __lowerCamelCase ( self ) -> Optional[Any]: # Check the shards when the dataset is a round multiple of total batch size. __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase , lowercase ) __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [0, 1, 2]], ] self.check_batch_sampler_shards(lowercase , lowercase ) __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 0, 1]], ] self.check_batch_sampler_shards(lowercase , lowercase ) __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __UpperCamelCase = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [1, 2, 3]], ] self.check_batch_sampler_shards(lowercase , lowercase ) __UpperCamelCase = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is very small. __UpperCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowercase , lowercase ) __UpperCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [[], []] self.check_batch_sampler_shards(lowercase , lowercase ) def __lowerCamelCase ( self ) -> Dict: # Check the shards when the dataset is a round multiple of batch size. __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size. __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [0, 1]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 0]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [1, 2]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is very small. __UpperCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) def __lowerCamelCase ( self ) -> Optional[Any]: # Check the shards when the dataset is a round multiple of total batch size. __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __UpperCamelCase = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is very small. __UpperCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) __UpperCamelCase = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) def __lowerCamelCase ( self ) -> str: # Check the shards when the dataset is a round multiple of batch size. __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size. __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is very small. __UpperCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) __UpperCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) def __lowerCamelCase ( self ) -> str: __UpperCamelCase = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 1_0, 1_1], [1_2, 1_3]] __UpperCamelCase = [BatchSamplerShard(lowercase , 2 , lowercase , even_batches=lowercase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [1_2, 1_3]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 1_0, 1_1]] ) def __lowerCamelCase ( self , lowercase , lowercase , lowercase , lowercase=False , lowercase=2 , lowercase=False ) -> List[str]: random.seed(lowercase ) __UpperCamelCase = list(lowercase ) __UpperCamelCase = [ IterableDatasetShard( lowercase , batch_size=lowercase , drop_last=lowercase , num_processes=lowercase , process_index=lowercase , split_batches=lowercase , ) for i in range(lowercase ) ] __UpperCamelCase = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowercase ) iterable_dataset_lists.append(list(lowercase ) ) __UpperCamelCase = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size __UpperCamelCase = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowercase ) , len(lowercase ) ) self.assertTrue(len(lowercase ) % shard_batch_size == 0 ) __UpperCamelCase = [] for idx in range(0 , len(lowercase ) , lowercase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowercase ) < len(lowercase ): reference += reference self.assertListEqual(lowercase , reference[: len(lowercase )] ) def __lowerCamelCase ( self ) -> Dict: __UpperCamelCase = 4_2 __UpperCamelCase = RandomIterableDataset() self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) # Edge case with a very small dataset __UpperCamelCase = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) def __lowerCamelCase ( self ) -> List[Any]: __UpperCamelCase = BatchSampler(range(1_6 ) , batch_size=4 , drop_last=lowercase ) __UpperCamelCase = SkipBatchSampler(lowercase , 2 ) self.assertListEqual(list(lowercase ) , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def __lowerCamelCase ( self ) -> List[Any]: __UpperCamelCase = SkipDataLoader(list(range(1_6 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def __lowerCamelCase ( self ) -> List[Any]: __UpperCamelCase = DataLoader(list(range(1_6 ) ) , batch_size=4 ) __UpperCamelCase = skip_first_batches(lowercase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def __lowerCamelCase ( self ) -> Tuple: __UpperCamelCase = DataLoaderShard(list(range(1_6 ) ) , batch_size=4 ) for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def __lowerCamelCase ( self ) -> Tuple: Accelerator() __UpperCamelCase = DataLoaderDispatcher(range(1_6 ) , batch_size=4 ) for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
243
1
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split UpperCamelCase = datasets.load_iris() UpperCamelCase = np.array(data['''data''']) UpperCamelCase = np.array(data['''target''']) UpperCamelCase = data['''target_names'''] UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = train_test_split(X, y) def lowercase_ ( _lowerCamelCase : str , _lowerCamelCase : Optional[Any]): return np.linalg.norm(np.array(_lowerCamelCase) - np.array(_lowerCamelCase)) def lowercase_ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Tuple , _lowerCamelCase : Any=5): lowercase__ : int = zip(_lowerCamelCase , _lowerCamelCase) # List of distances of all points from the point to be classified lowercase__ : str = [] for data_point in data: lowercase__ : str = euclidean_distance(data_point[0] , _lowerCamelCase) distances.append((distance, data_point[1])) # Choosing 'k' points with the least distances. lowercase__ : Dict = [i[1] for i in sorted(_lowerCamelCase)[:k]] # Most commonly occurring class among them # is the class into which the point is classified lowercase__ : Tuple = Counter(_lowerCamelCase).most_common(1)[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
87
def a( A : list ) -> list: """simple docstring""" if any(not isinstance(A , A ) or x < 0 for x in sequence ): raise TypeError("Sequence must be list of non-negative integers" ) for _ in range(len(A ) ): for i, (rod_upper, rod_lower) in enumerate(zip(A , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
227
0
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCAmelCase ( __a ): snake_case : str = ["""image_processor""", """tokenizer"""] snake_case : int = """LayoutLMv2ImageProcessor""" snake_case : Any = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__(self , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ ): if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , lowerCAmelCase__ , ) _UpperCAmelCase : Any = kwargs.pop("""feature_extractor""" ) _UpperCAmelCase : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(lowerCAmelCase__ , lowerCAmelCase__ ) def __call__(self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 0 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = True , lowerCAmelCase__ = None , **lowerCAmelCase__ , ): # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( """You cannot provide bounding boxes """ """if you initialized the image processor with apply_ocr set to True.""" ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( """You cannot provide word labels if you initialized the image processor with apply_ocr set to True.""" ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError("""You cannot return overflowing tokens without returning the offsets mapping.""" ) # first, apply the image processor _UpperCAmelCase : List[str] = self.image_processor(images=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : int = [text] # add batch dimension (as the image processor always adds a batch dimension) _UpperCAmelCase : Union[str, Any] = features["""words"""] _UpperCAmelCase : Union[str, Any] = self.tokenizer( text=text if text is not None else features["""words"""] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["""boxes"""] , word_labels=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) # add pixel values _UpperCAmelCase : Union[str, Any] = features.pop("""pixel_values""" ) if return_overflowing_tokens is True: _UpperCAmelCase : Dict = self.get_overflowing_images(lowerCAmelCase__ , encoded_inputs["""overflow_to_sample_mapping"""] ) _UpperCAmelCase : str = images return encoded_inputs def snake_case_ (self , lowerCAmelCase__ , lowerCAmelCase__ ): # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image _UpperCAmelCase : Tuple = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(lowerCAmelCase__ ) != len(lowerCAmelCase__ ): raise ValueError( """Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got""" F" {len(lowerCAmelCase__ )} and {len(lowerCAmelCase__ )}" ) return images_with_overflow def snake_case_ (self , *lowerCAmelCase__ , **lowerCAmelCase__ ): return self.tokenizer.batch_decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) def snake_case_ (self , *lowerCAmelCase__ , **lowerCAmelCase__ ): return self.tokenizer.decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) @property def snake_case_ (self ): return ["input_ids", "bbox", "attention_mask", "image"] @property def snake_case_ (self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , lowerCAmelCase__ , ) return self.image_processor_class @property def snake_case_ (self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , lowerCAmelCase__ , ) return self.image_processor
170
'''simple docstring''' import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline lowerCAmelCase_ : Any = datasets.utils.logging.get_logger(__name__) @dataclass class __lowerCAmelCase ( datasets.BuilderConfig ): snake_case : Optional[datasets.Features] = None snake_case : str = "utf-8" snake_case : Optional[str] = None snake_case : Optional[str] = None snake_case : bool = True # deprecated snake_case : Optional[int] = None # deprecated snake_case : int = 1_0 << 2_0 # 10MB snake_case : Optional[bool] = None class __lowerCAmelCase ( datasets.ArrowBasedBuilder ): snake_case : Optional[int] = JsonConfig def snake_case_ (self ): if self.config.block_size is not None: logger.warning("""The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead""" ) _UpperCAmelCase : List[Any] = self.config.block_size if self.config.use_threads is not True: logger.warning( """The JSON loader parameter `use_threads` is deprecated and doesn't have any effect anymore.""" ) if self.config.newlines_in_values is not None: raise ValueError("""The JSON loader parameter `newlines_in_values` is no longer supported""" ) return datasets.DatasetInfo(features=self.config.features ) def snake_case_ (self , lowerCAmelCase__ ): if not self.config.data_files: raise ValueError(F"At least one data file must be specified, but got data_files={self.config.data_files}" ) _UpperCAmelCase : Tuple = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): _UpperCAmelCase : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : int = [files] _UpperCAmelCase : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] _UpperCAmelCase : List[Any] = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : Tuple = [files] _UpperCAmelCase : Union[str, Any] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"""files""": files} ) ) return splits def snake_case_ (self , lowerCAmelCase__ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): _UpperCAmelCase : Tuple = self.config.features.arrow_schema.field(lowerCAmelCase__ ).type _UpperCAmelCase : Any = pa_table.append_column(lowerCAmelCase__ , pa.array([None] * len(lowerCAmelCase__ ) , type=lowerCAmelCase__ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example _UpperCAmelCase : Optional[int] = table_cast(lowerCAmelCase__ , self.config.features.arrow_schema ) return pa_table def snake_case_ (self , lowerCAmelCase__ ): for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(lowerCAmelCase__ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: _UpperCAmelCase : Optional[Any] = json.load(lowerCAmelCase__ ) # We keep only the field we are interested in _UpperCAmelCase : Tuple = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(lowerCAmelCase__ , (list, tuple) ): _UpperCAmelCase : Union[str, Any] = set().union(*[row.keys() for row in dataset] ) _UpperCAmelCase : Optional[int] = {col: [row.get(lowerCAmelCase__ ) for row in dataset] for col in keys} else: _UpperCAmelCase : Optional[int] = dataset _UpperCAmelCase : Any = pa.Table.from_pydict(lowerCAmelCase__ ) yield file_idx, self._cast_table(lowerCAmelCase__ ) # If the file has one json object per line else: with open(lowerCAmelCase__ , """rb""" ) as f: _UpperCAmelCase : Optional[int] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small _UpperCAmelCase : List[Any] = max(self.config.chunksize // 3_2 , 1_6 << 1_0 ) _UpperCAmelCase : Union[str, Any] = ( self.config.encoding_errors if self.config.encoding_errors is not None else """strict""" ) while True: _UpperCAmelCase : List[str] = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(lowerCAmelCase__ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": _UpperCAmelCase : int = batch.decode(self.config.encoding , errors=lowerCAmelCase__ ).encode("""utf-8""" ) try: while True: try: _UpperCAmelCase : Tuple = paj.read_json( io.BytesIO(lowerCAmelCase__ ) , read_options=paj.ReadOptions(block_size=lowerCAmelCase__ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(lowerCAmelCase__ , pa.ArrowInvalid ) and "straddling" not in str(lowerCAmelCase__ ) or block_size > len(lowerCAmelCase__ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"Batch of {len(lowerCAmelCase__ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}." ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( lowerCAmelCase__ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: _UpperCAmelCase : Optional[Any] = json.load(lowerCAmelCase__ ) except json.JSONDecodeError: logger.error(F"Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # list is the only sequence type supported in JSON try: _UpperCAmelCase : Tuple = set().union(*[row.keys() for row in dataset] ) _UpperCAmelCase : Dict = {col: [row.get(lowerCAmelCase__ ) for row in dataset] for col in keys} _UpperCAmelCase : int = pa.Table.from_pydict(lowerCAmelCase__ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}" ) raise ValueError(F"Not able to read records in the JSON file at {file}." ) from None yield file_idx, self._cast_table(lowerCAmelCase__ ) break else: logger.error(F"Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}" ) raise ValueError( F"Not able to read records in the JSON file at {file}. " F"You should probably indicate the field of the JSON file containing your records. " F"This JSON file contain the following fields: {str(list(dataset.keys() ) )}. " F"Select the correct one and provide it as `field='XXX'` to the dataset loading method. " ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) batch_idx += 1
170
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_: Dict =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Union[str, Any] ={ 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/config.json', } class __A ( UpperCamelCase__ ): a__ : List[str] = """gpt_neox_japanese""" def __init__(self : List[Any] , __a : str=32000 , __a : int=2560 , __a : Dict=32 , __a : List[str]=32 , __a : str=4 , __a : List[str]="gelu" , __a : List[str]=1.00 , __a : Union[str, Any]=10000 , __a : str=2048 , __a : Tuple=0.02 , __a : List[Any]=1E-5 , __a : List[str]=True , __a : str=31996 , __a : Optional[Any]=31999 , __a : Any=0.1 , __a : List[Any]=0.0 , **__a : str , ): super().__init__(bos_token_id=__a , eos_token_id=__a , **__a ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_multiple_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = rotary_pct UpperCAmelCase_ = rotary_emb_base UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = use_cache UpperCAmelCase_ = attention_dropout UpperCAmelCase_ = hidden_dropout
1
from __future__ import annotations def snake_case_ ( lowerCAmelCase_ : int , lowerCAmelCase_ : int ): if partitions <= 0: raise ValueError("""partitions must be a positive number!""" ) if partitions > number_of_bytes: raise ValueError("""partitions can not > number_of_bytes!""" ) __lowercase : Dict = number_of_bytes // partitions __lowercase : Union[str, Any] = [] for i in range(lowerCAmelCase_ ): __lowercase : str = i * bytes_per_partition + 1 __lowercase : List[Any] = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F"{start_bytes}-{end_bytes}" ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
233
0
'''simple docstring''' import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : Tuple , UpperCamelCase__ : UNetaDModel , UpperCamelCase__ : UNetaDModel , UpperCamelCase__ : DDPMScheduler , UpperCamelCase__ : Dict , ): """simple docstring""" super().__init__() UpperCamelCase = value_function UpperCamelCase = unet UpperCamelCase = scheduler UpperCamelCase = env UpperCamelCase = env.get_dataset() UpperCamelCase = {} for key in self.data.keys(): try: UpperCamelCase = self.data[key].mean() except: # noqa: E722 pass UpperCamelCase = {} for key in self.data.keys(): try: UpperCamelCase = self.data[key].std() except: # noqa: E722 pass UpperCamelCase = env.observation_space.shape[0] UpperCamelCase = env.action_space.shape[0] def A ( self : Optional[Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any ): """simple docstring""" return (x_in - self.means[key]) / self.stds[key] def A ( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ): """simple docstring""" return x_in * self.stds[key] + self.means[key] def A ( self : str , UpperCamelCase__ : str ): """simple docstring""" if type(UpperCamelCase__ ) is dict: return {k: self.to_torch(UpperCamelCase__ ) for k, v in x_in.items()} elif torch.is_tensor(UpperCamelCase__ ): return x_in.to(self.unet.device ) return torch.tensor(UpperCamelCase__ , device=self.unet.device ) def A ( self : Dict , UpperCamelCase__ : List[str] , UpperCamelCase__ : str , UpperCamelCase__ : List[str] ): """simple docstring""" for key, val in cond.items(): UpperCamelCase = val.clone() return x_in def A ( self : str , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any] ): """simple docstring""" UpperCamelCase = x.shape[0] UpperCamelCase = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model UpperCamelCase = torch.full((batch_size,) , UpperCamelCase__ , device=self.unet.device , dtype=torch.long ) for _ in range(UpperCamelCase__ ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models UpperCamelCase = self.value_function(x.permute(0 , 2 , 1 ) , UpperCamelCase__ ).sample UpperCamelCase = torch.autograd.grad([y.sum()] , [x] )[0] UpperCamelCase = self.scheduler._get_variance(UpperCamelCase__ ) UpperCamelCase = torch.exp(0.5 * posterior_variance ) UpperCamelCase = model_std * grad UpperCamelCase = 0 UpperCamelCase = x.detach() UpperCamelCase = x + scale * grad UpperCamelCase = self.reset_xa(UpperCamelCase__ , UpperCamelCase__ , self.action_dim ) UpperCamelCase = self.unet(x.permute(0 , 2 , 1 ) , UpperCamelCase__ ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg UpperCamelCase = self.scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , predict_epsilon=UpperCamelCase__ )['prev_sample'] # apply conditions to the trajectory (set the initial state) UpperCamelCase = self.reset_xa(UpperCamelCase__ , UpperCamelCase__ , self.action_dim ) UpperCamelCase = self.to_torch(UpperCamelCase__ ) return x, y def __call__( self : List[Any] , UpperCamelCase__ : Dict , UpperCamelCase__ : str=6_4 , UpperCamelCase__ : List[str]=3_2 , UpperCamelCase__ : Union[str, Any]=2 , UpperCamelCase__ : Tuple=0.1 ): """simple docstring""" UpperCamelCase = self.normalize(UpperCamelCase__ , 'observations' ) UpperCamelCase = obs[None].repeat(UpperCamelCase__ , axis=0 ) UpperCamelCase = {0: self.to_torch(UpperCamelCase__ )} UpperCamelCase = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) UpperCamelCase = randn_tensor(UpperCamelCase__ , device=self.unet.device ) UpperCamelCase = self.reset_xa(UpperCamelCase__ , UpperCamelCase__ , self.action_dim ) UpperCamelCase = self.to_torch(UpperCamelCase__ ) # run the diffusion process UpperCamelCase , UpperCamelCase = self.run_diffusion(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # sort output trajectories by value UpperCamelCase = y.argsort(0 , descending=UpperCamelCase__ ).squeeze() UpperCamelCase = x[sorted_idx] UpperCamelCase = sorted_values[:, :, : self.action_dim] UpperCamelCase = actions.detach().cpu().numpy() UpperCamelCase = self.de_normalize(UpperCamelCase__ , key='actions' ) # select the action with the highest value if y is not None: UpperCamelCase = 0 else: # if we didn't run value guiding, select a random action UpperCamelCase = np.random.randint(0 , UpperCamelCase__ ) UpperCamelCase = denorm_actions[selected_index, 0] return denorm_actions
249
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _lowerCamelCase : Any = {"configuration_vit_mae": ["VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMAEConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : int = [ "VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMAEForPreTraining", "ViTMAELayer", "ViTMAEModel", "ViTMAEPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Union[str, Any] = [ "TFViTMAEForPreTraining", "TFViTMAEModel", "TFViTMAEPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys _lowerCamelCase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
249
1
"""simple docstring""" from __future__ import annotations A_ = '''Muhammad Umer Farooq''' A_ = '''MIT''' A_ = '''1.0.0''' A_ = '''Muhammad Umer Farooq''' A_ = '''contact@muhammadumerfarooq.me''' A_ = '''Alpha''' import re from html.parser import HTMLParser from urllib import parse import requests class lowercase( __a ): '''simple docstring''' def __init__( self: Optional[Any], a_: str ): '''simple docstring''' super().__init__() _snake_case : list[str] = [] _snake_case : List[Any] = domain def UpperCamelCase_ ( self: List[Any], a_: str, a_: list[tuple[str, str | None]] ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: _snake_case : Any = parse.urljoin(self.domain, a_ ) self.urls.append(a_ ) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" return ".".join(get_sub_domain_name(snake_case__ ).split(""".""" )[-2:] ) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" return parse.urlparse(snake_case__ ).netloc def UpperCAmelCase__ (snake_case__ : str = "https://github.com" ): """simple docstring""" _snake_case : Dict = get_domain_name(snake_case__ ) # Initialize the parser _snake_case : List[str] = Parser(snake_case__ ) try: # Open URL _snake_case : Union[str, Any] = requests.get(snake_case__ ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through _snake_case : str = set() for link in parser.urls: # open URL. # read = requests.get(link) try: _snake_case : Optional[Any] = requests.get(snake_case__ ) # Get the valid email. _snake_case : Tuple = re.findall("""[a-zA-Z0-9]+@""" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(snake_case__ ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(snake_case__ ) if __name__ == "__main__": A_ = emails_from_url('''https://github.com''') print(F'''{len(emails)} emails found:''') print('''\n'''.join(sorted(emails)))
64
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } lowerCAmelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } lowerCAmelCase__ = '</w>' lowerCAmelCase__ = '@@ ' def _UpperCAmelCase (UpperCamelCase__ : Optional[Any] ): _A : Optional[int] = set() _A : Optional[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _A : List[Any] = char return pairs # Speech2Text2 has no max input length lowerCAmelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 10_24} class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] def __init__( self , __lowerCamelCase , __lowerCamelCase="<s>" , __lowerCamelCase="<pad>" , __lowerCamelCase="</s>" , __lowerCamelCase="<unk>" , __lowerCamelCase=False , __lowerCamelCase=None , **__lowerCamelCase , ) -> Optional[Any]: super().__init__( unk_token=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , pad_token=__lowerCamelCase , do_lower_case=__lowerCamelCase , **__lowerCamelCase , ) _A : Dict = do_lower_case with open(__lowerCamelCase , encoding="utf-8") as vocab_handle: _A : Optional[int] = json.load(__lowerCamelCase) _A : Optional[Any] = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(F"No merges files provided. {self.__class__.__name__} can only be used for decoding.") _A : Optional[Any] = None _A : Tuple = None else: with open(__lowerCamelCase , encoding="utf-8") as merges_handle: _A : Optional[int] = merges_handle.read().split("\n")[:-1] _A : Union[str, Any] = [tuple(merge.split()[:2]) for merge in merges] _A : Optional[int] = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase)))) _A : List[Any] = {} @property def _lowerCamelCase ( self) -> int: return len(self.decoder) def _lowerCamelCase ( self) -> Dict: return dict(self.encoder , **self.added_tokens_encoder) def _lowerCamelCase ( self , __lowerCamelCase) -> Dict: _A : Tuple = tuple(token[:-1]) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] _A : int = get_pairs(__lowerCamelCase) if not pairs: return token while True: _A : Any = min(__lowerCamelCase , key=lambda __lowerCamelCase: self.bpe_ranks.get(__lowerCamelCase , float("inf"))) if bigram not in self.bpe_ranks: break _A , _A : Optional[int] = bigram _A : int = [] _A : str = 0 while i < len(__lowerCamelCase): try: _A : str = word.index(__lowerCamelCase , __lowerCamelCase) except ValueError: new_word.extend(word[i:]) break else: new_word.extend(word[i:j]) _A : str = j if word[i] == first and i < len(__lowerCamelCase) - 1 and word[i + 1] == second: new_word.append(first + second) i += 2 else: new_word.append(word[i]) i += 1 _A : List[str] = tuple(__lowerCamelCase) _A : List[str] = new_word if len(__lowerCamelCase) == 1: break else: _A : List[Any] = get_pairs(__lowerCamelCase) _A : Tuple = " ".join(__lowerCamelCase) if word == "\n " + BPE_TOKEN_MERGES: _A : List[str] = "\n" + BPE_TOKEN_MERGES if word.endswith(__lowerCamelCase): _A : int = word.replace(__lowerCamelCase , "") _A : int = word.replace(" " , __lowerCamelCase) _A : Union[str, Any] = word return word def _lowerCamelCase ( self , __lowerCamelCase) -> Optional[Any]: if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding.") if self.do_lower_case: _A : List[Any] = text.lower() _A : Optional[int] = text.split() _A : List[str] = [] for token in text: if token: split_tokens.extend(list(self.bpe(__lowerCamelCase).split(" "))) return split_tokens def _lowerCamelCase ( self , __lowerCamelCase) -> int: return self.encoder.get(__lowerCamelCase , self.encoder.get(self.unk_token)) def _lowerCamelCase ( self , __lowerCamelCase) -> str: _A : List[str] = self.decoder.get(__lowerCamelCase , self.unk_token) return result def _lowerCamelCase ( self , __lowerCamelCase) -> str: _A : str = " ".join(__lowerCamelCase) # make sure @@ tokens are concatenated _A : int = "".join(string.split(__lowerCamelCase)) return string def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> Tuple[str]: if not os.path.isdir(__lowerCamelCase): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return _A : Any = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) _A : Any = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"]) with open(__lowerCamelCase , "w" , encoding="utf-8") as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__lowerCamelCase , ensure_ascii=__lowerCamelCase) + "\n") _A : Union[str, Any] = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__lowerCamelCase , "w" , encoding="utf-8") as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __lowerCamelCase: kv[1]): if index != token_index: logger.warning( F"Saving vocabulary to {merges_file}: BPE merge indices are not consecutive." " Please check that the tokenizer is not corrupted!") _A : Optional[int] = token_index writer.write(" ".join(__lowerCamelCase) + "\n") index += 1 return (vocab_file, merges_file)
11
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = args.pruning_method UpperCAmelCase = args.threshold UpperCAmelCase = args.model_name_or_path.rstrip("""/""" ) UpperCAmelCase = args.target_model_path print(F'''Load fine-pruned model from {model_name_or_path}''' ) UpperCAmelCase = torch.load(os.path.join(_snake_case , """pytorch_model.bin""" ) ) UpperCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) elif "bias" in name: UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) else: if pruning_method == "magnitude": UpperCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue UpperCAmelCase = name[:-6] UpperCAmelCase = model[F'''{prefix_}mask_scores'''] UpperCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue UpperCAmelCase = name[:-6] UpperCAmelCase = model[F'''{prefix_}mask_scores'''] UpperCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue UpperCAmelCase = name[:-6] UpperCAmelCase = model[F'''{prefix_}mask_scores'''] UpperCAmelCase , UpperCAmelCase = -0.1, 1.1 UpperCAmelCase = torch.sigmoid(_snake_case ) UpperCAmelCase = s * (r - l) + l UpperCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) else: raise ValueError("""Unknown pruning method""" ) if target_model_path is None: UpperCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F'''bertarized_{os.path.basename(_snake_case )}''' ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F'''\nCreated folder {target_model_path}''' ) torch.save(_snake_case , os.path.join(_snake_case , """pytorch_model.bin""" ) ) print("""\nPruned model saved! See you later!""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) _UpperCamelCase = parser.parse_args() main(args)
234
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _UpperCamelCase = None _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = """▁""" _UpperCamelCase = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCamelCase = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } _UpperCamelCase = { """google/pegasus-xsum""": 512, } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = PegasusTokenizer SCREAMING_SNAKE_CASE = ['''input_ids''', '''attention_mask'''] def __init__( self ,A=None ,A=None ,A="<pad>" ,A="</s>" ,A="<unk>" ,A="<mask_2>" ,A="<mask_1>" ,A=None ,A=103 ,**A ,): UpperCAmelCase = offset if additional_special_tokens is not None: if not isinstance(A ,A ): raise TypeError( F'''additional_special_tokens should be of type {type(A )}, but is''' F''' {type(A )}''' ) UpperCAmelCase = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'''<unk_{i}>''' for i in range(len(A ) ,self.offset - 1 ) ] if len(set(A ) ) != len(A ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" F''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) UpperCAmelCase = additional_special_tokens_extended else: UpperCAmelCase = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'''<unk_{i}>''' for i in range(2 ,self.offset )] super().__init__( A ,tokenizer_file=A ,pad_token=A ,eos_token=A ,unk_token=A ,mask_token=A ,mask_token_sent=A ,offset=A ,additional_special_tokens=A ,**A ,) UpperCAmelCase = vocab_file UpperCAmelCase = False if not self.vocab_file else True def _UpperCamelCase ( self ,A ): UpperCAmelCase = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( """There should be 3 special tokens: mask_token, pad_token, and eos_token +""" F''' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}''' ) return [1 if x in all_special_ids else 0 for x in seq] def _UpperCamelCase ( self ,A ,A = None ,A = False ): if already_has_special_tokens: return self._special_token_mask(A ) elif token_ids_a is None: return self._special_token_mask(A ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _UpperCamelCase ( self ,A ,A=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _UpperCamelCase ( self ,A ,A = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(A ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ): copyfile(self.vocab_file ,A ) return (out_vocab_file,)
234
1
"""simple docstring""" _a = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _a = [{'type': 'code', 'content': INSTALL_CONTENT}] _a = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
17
from math import ceil def A ( _lowercase = 1_001 ): SCREAMING_SNAKE_CASE : Any = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): SCREAMING_SNAKE_CASE : Union[str, Any] = 2 * i + 1 SCREAMING_SNAKE_CASE : int = 2 * i SCREAMING_SNAKE_CASE : List[str] = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: __UpperCamelCase : Dict = int(sys.argv[1]) print(solution(n)) except ValueError: print('Invalid entry - please enter a number')
182
0
def __A ( _lowercase , _lowercase ): '''simple docstring''' return int(input_a == input_a == 0 ) def __A ( ): '''simple docstring''' print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(f"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(f"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(f"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
364
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class SCREAMING_SNAKE_CASE ( snake_case ): """simple docstring""" @require_torch def __A ( self: Dict ) -> Optional[int]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _A = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(__A ) BertModel.from_pretrained(__A ) BertTokenizer.from_pretrained(__A ) pipeline(task='''fill-mask''' , model=__A ) # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _A = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Dict ) -> Tuple: # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _A = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(__A ) BertModel.from_pretrained(__A ) BertTokenizer.from_pretrained(__A ) pipeline(task='''fill-mask''' , model=__A ) # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Any ) -> Optional[Any]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert-sharded" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled") socket.socket = offline_socket ''' # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # next emulate no network _A = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Optional[int] ) -> Dict: _A = ''' from transformers import pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" pipe = pipeline(model=mname) ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled") socket.socket = offline_socket ''' _A = self.get_env() _A = '''1''' _A = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( '''You cannot infer task automatically within `pipeline` when using offline mode''' , result.stderr.decode().replace('''\n''' , '''''' ) , ) @require_torch def __A ( self: Optional[int] ) -> int: _A = ''' from transformers import AutoModel ''' _A = ''' mname = "hf-internal-testing/test_dynamic_model" AutoModel.from_pretrained(mname, trust_remote_code=True) print("success") ''' # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() )
75
0
import os import sys import unittest _A = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path _A = os.path.join(git_repo_path, "src", "transformers") _A = "\n{0} = None\n" _A = "\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n" _A = "\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n" class _lowerCAmelCase ( unittest.TestCase ): def __a ( self ) -> List[Any]: lowerCAmelCase_ = find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(_UpperCamelCase ) lowerCAmelCase_ = find_backend(" if not is_tokenizers_available():" ) self.assertEqual(_UpperCamelCase , "tokenizers" ) lowerCAmelCase_ = find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(_UpperCamelCase , "tensorflow_text" ) lowerCAmelCase_ = find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tokenizers" ) lowerCAmelCase_ = find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tensorflow_text" ) lowerCAmelCase_ = find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tokenizers_and_vision" ) def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch" , _UpperCamelCase ) self.assertIn("tensorflow_text" , _UpperCamelCase ) self.assertIn("sentencepiece_and_tokenizers" , _UpperCamelCase ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertModel" , objects["tf"] ) self.assertIn("FlaxBertModel" , objects["flax"] ) self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertTokenizer" , objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer" , objects["sentencepiece_and_tokenizers"] ) def __a ( self ) -> List[str]: lowerCAmelCase_ = create_dummy_object("CONSTANT" , "'torch'" ) self.assertEqual(_UpperCamelCase , "\nCONSTANT = None\n" ) lowerCAmelCase_ = create_dummy_object("function" , "'torch'" ) self.assertEqual( _UpperCamelCase , "\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n" ) lowerCAmelCase_ = "\nclass FakeClass(metaclass=DummyObject):\n _backends = 'torch'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, 'torch')\n" lowerCAmelCase_ = create_dummy_object("FakeClass" , "'torch'" ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def __a ( self ) -> str: lowerCAmelCase_ = "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" lowerCAmelCase_ = create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"] , _UpperCamelCase )
231
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class _lowerCAmelCase : def __init__( self , _UpperCamelCase=2 , _UpperCamelCase=3 , _UpperCamelCase=64 , _UpperCamelCase=None ) -> Optional[Any]: lowerCAmelCase_ = np.random.default_rng(_UpperCamelCase ) lowerCAmelCase_ = length lowerCAmelCase_ = rng.normal(size=(length,) ).astype(np.floataa ) lowerCAmelCase_ = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> int: return self.length def __getitem__( self , _UpperCamelCase ) -> Dict: return {"x": self.x[i], "y": self.y[i]} class _lowerCAmelCase ( torch.nn.Module ): def __init__( self , _UpperCamelCase=0 , _UpperCamelCase=0 , _UpperCamelCase=False ) -> List[Any]: super().__init__() lowerCAmelCase_ = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase_ = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase_ = True def __a ( self , _UpperCamelCase=None ) -> Any: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) lowerCAmelCase_ = False return x * self.a[0] + self.b[0] class _lowerCAmelCase ( torch.nn.Module ): def __init__( self , _UpperCamelCase=0 , _UpperCamelCase=0 , _UpperCamelCase=False ) -> Optional[int]: super().__init__() lowerCAmelCase_ = torch.nn.Parameter(torch.tensor(_UpperCamelCase ).float() ) lowerCAmelCase_ = torch.nn.Parameter(torch.tensor(_UpperCamelCase ).float() ) lowerCAmelCase_ = True def __a ( self , _UpperCamelCase=None ) -> int: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) lowerCAmelCase_ = False return x * self.a + self.b def lowerCamelCase__ ( __lowerCAmelCase : List[Any] , __lowerCAmelCase : int = 16 ): """simple docstring""" from datasets import load_dataset from transformers import AutoTokenizer lowerCAmelCase_ = AutoTokenizer.from_pretrained("bert-base-cased" ) lowerCAmelCase_ = {"train": "tests/test_samples/MRPC/train.csv", "validation": "tests/test_samples/MRPC/dev.csv"} lowerCAmelCase_ = load_dataset("csv" , data_files=__lowerCAmelCase ) lowerCAmelCase_ = datasets["train"].unique("label" ) lowerCAmelCase_ = {v: i for i, v in enumerate(__lowerCAmelCase )} def tokenize_function(__lowerCAmelCase : List[Any] ): # max_length=None => use the model max length (it's actually the default) lowerCAmelCase_ = tokenizer( examples["sentence1"] , examples["sentence2"] , truncation=__lowerCAmelCase , max_length=__lowerCAmelCase , padding="max_length" ) if "label" in examples: lowerCAmelCase_ = [label_to_id[l] for l in examples["label"]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowerCAmelCase_ = datasets.map( __lowerCAmelCase , batched=__lowerCAmelCase , remove_columns=["sentence1", "sentence2", "label"] , ) def collate_fn(__lowerCAmelCase : Optional[int] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(__lowerCAmelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return tokenizer.pad(__lowerCAmelCase , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. lowerCAmelCase_ = DataLoader(tokenized_datasets["train"] , shuffle=__lowerCAmelCase , collate_fn=__lowerCAmelCase , batch_size=2 ) lowerCAmelCase_ = DataLoader(tokenized_datasets["validation"] , shuffle=__lowerCAmelCase , collate_fn=__lowerCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
231
1
'''simple docstring''' import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = MobileBertConfig.from_json_file(lowerCAmelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _UpperCAmelCase : Any = MobileBertForPreTraining(lowerCAmelCase_ ) # Load weights from tf checkpoint _UpperCAmelCase : Optional[int] = load_tf_weights_in_mobilebert(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , lowerCAmelCase_ ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A_ : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
349
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ : List[Any] = logging.get_logger(__name__) A_ : Union[str, Any] = { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json""" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """roformer""" def __init__( self ,a_=50_000 ,a_=None ,a_=768 ,a_=12 ,a_=12 ,a_=3_072 ,a_="gelu" ,a_=0.1 ,a_=0.1 ,a_=1_536 ,a_=2 ,a_=0.02 ,a_=1E-1_2 ,a_=0 ,a_=False ,a_=True ,**a_ ,) -> Tuple: super().__init__(pad_token_id=a_ ,**a_ ) _UpperCAmelCase : List[Any] = vocab_size _UpperCAmelCase : str = hidden_size if embedding_size is None else embedding_size _UpperCAmelCase : List[Any] = hidden_size _UpperCAmelCase : str = num_hidden_layers _UpperCAmelCase : Optional[Any] = num_attention_heads _UpperCAmelCase : Optional[Any] = hidden_act _UpperCAmelCase : str = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Any = attention_probs_dropout_prob _UpperCAmelCase : Optional[int] = max_position_embeddings _UpperCAmelCase : Any = type_vocab_size _UpperCAmelCase : Tuple = initializer_range _UpperCAmelCase : Dict = layer_norm_eps _UpperCAmelCase : Optional[int] = rotary_value _UpperCAmelCase : Any = use_cache class lowercase ( _lowerCamelCase ): """simple docstring""" @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCAmelCase : Optional[Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _UpperCAmelCase : List[Any] = {0: """batch""", 1: """sequence"""} _UpperCAmelCase : Tuple = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
349
1
"""simple docstring""" from collections import deque from math import floor from random import random from time import time class UpperCAmelCase_ : def __init__( self ) -> Tuple: __lowercase : Dict = {} def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=1 ) -> Tuple: if self.graph.get(UpperCamelCase_ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: __lowercase : Optional[int] = [[w, v]] if not self.graph.get(UpperCamelCase_ ): __lowercase : Tuple = [] def _lowerCamelCase ( self ) -> int: return list(self.graph ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ) -> Tuple: if self.graph.get(UpperCamelCase_ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(UpperCamelCase_ ) def _lowerCamelCase ( self , UpperCamelCase_=-2 , UpperCamelCase_=-1 ) -> Optional[int]: if s == d: return [] __lowercase : Optional[int] = [] __lowercase : Optional[int] = [] if s == -2: __lowercase : Dict = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Optional[int] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : Optional[int] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(UpperCamelCase_ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) __lowercase : List[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(UpperCamelCase_ ) != 0: __lowercase : List[Any] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : str = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return visited def _lowerCamelCase ( self , UpperCamelCase_=-1 ) -> List[str]: if c == -1: __lowercase : Union[str, Any] = floor(random() * 1_00_00 ) + 10 for i in range(UpperCamelCase_ ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): __lowercase : Optional[int] = floor(random() * c ) + 1 if n != i: self.add_pair(UpperCamelCase_ , UpperCamelCase_ , 1 ) def _lowerCamelCase ( self , UpperCamelCase_=-2 ) -> Dict: __lowercase : Dict = deque() __lowercase : Tuple = [] if s == -2: __lowercase : Tuple = list(self.graph )[0] d.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) while d: __lowercase : Tuple = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _lowerCamelCase ( self , UpperCamelCase_ ) -> int: __lowercase : int = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _lowerCamelCase ( self , UpperCamelCase_ ) -> Union[str, Any]: return len(self.graph[u] ) def _lowerCamelCase ( self , UpperCamelCase_=-2 ) -> Dict: __lowercase : Dict = [] __lowercase : Optional[int] = [] if s == -2: __lowercase : List[Any] = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Optional[Any] = s __lowercase : Optional[Any] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) __lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(UpperCamelCase_ ) != 0: __lowercase : Dict = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : int = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return sorted_nodes def _lowerCamelCase ( self ) -> int: __lowercase : Optional[int] = [] __lowercase : Optional[Any] = [] __lowercase : Dict = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Union[str, Any] = -2 __lowercase : Optional[int] = [] __lowercase : Optional[int] = s __lowercase : Tuple = False __lowercase : Union[str, Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): __lowercase : str = len(UpperCamelCase_ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) __lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() __lowercase : Union[str, Any] = True if len(UpperCamelCase_ ) != 0: __lowercase : Optional[Any] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : str = False indirect_parents.append(UpperCamelCase_ ) __lowercase : Tuple = s __lowercase : Dict = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return list(UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Optional[int]: __lowercase : Optional[int] = [] __lowercase : Union[str, Any] = [] __lowercase : str = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Optional[int] = -2 __lowercase : List[Any] = [] __lowercase : Tuple = s __lowercase : Any = False __lowercase : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): __lowercase : Optional[int] = len(UpperCamelCase_ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) __lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() __lowercase : List[str] = True if len(UpperCamelCase_ ) != 0: __lowercase : Optional[int] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : int = False indirect_parents.append(UpperCamelCase_ ) __lowercase : Tuple = s __lowercase : Optional[Any] = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return False def _lowerCamelCase ( self , UpperCamelCase_=-2 , UpperCamelCase_=-1 ) -> List[str]: __lowercase : int = time() self.dfs(UpperCamelCase_ , UpperCamelCase_ ) __lowercase : List[Any] = time() return end - begin def _lowerCamelCase ( self , UpperCamelCase_=-2 ) -> str: __lowercase : Dict = time() self.bfs(UpperCamelCase_ ) __lowercase : Dict = time() return end - begin class UpperCAmelCase_ : def __init__( self ) -> Optional[Any]: __lowercase : Optional[int] = {} def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=1 ) -> List[Any]: # check if the u exists if self.graph.get(UpperCamelCase_ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist __lowercase : Union[str, Any] = [[w, v]] # add the other way if self.graph.get(UpperCamelCase_ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist __lowercase : Union[str, Any] = [[w, u]] def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ ) -> Union[str, Any]: if self.graph.get(UpperCamelCase_ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(UpperCamelCase_ ) # the other way round if self.graph.get(UpperCamelCase_ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(UpperCamelCase_ ) def _lowerCamelCase ( self , UpperCamelCase_=-2 , UpperCamelCase_=-1 ) -> Optional[int]: if s == d: return [] __lowercase : Optional[Any] = [] __lowercase : str = [] if s == -2: __lowercase : Dict = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : List[Any] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : Tuple = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(UpperCamelCase_ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) __lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(UpperCamelCase_ ) != 0: __lowercase : Union[str, Any] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : List[str] = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return visited def _lowerCamelCase ( self , UpperCamelCase_=-1 ) -> Dict: if c == -1: __lowercase : Dict = floor(random() * 1_00_00 ) + 10 for i in range(UpperCamelCase_ ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): __lowercase : Dict = floor(random() * c ) + 1 if n != i: self.add_pair(UpperCamelCase_ , UpperCamelCase_ , 1 ) def _lowerCamelCase ( self , UpperCamelCase_=-2 ) -> int: __lowercase : Union[str, Any] = deque() __lowercase : Optional[Any] = [] if s == -2: __lowercase : Optional[Any] = list(self.graph )[0] d.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) while d: __lowercase : List[str] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _lowerCamelCase ( self , UpperCamelCase_ ) -> Dict: return len(self.graph[u] ) def _lowerCamelCase ( self ) -> Union[str, Any]: __lowercase : List[str] = [] __lowercase : int = [] __lowercase : Optional[int] = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Optional[Any] = -2 __lowercase : Optional[Any] = [] __lowercase : List[str] = s __lowercase : Any = False __lowercase : Tuple = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): __lowercase : Optional[int] = len(UpperCamelCase_ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) __lowercase : List[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() __lowercase : Tuple = True if len(UpperCamelCase_ ) != 0: __lowercase : List[Any] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : str = False indirect_parents.append(UpperCamelCase_ ) __lowercase : List[str] = s __lowercase : Optional[int] = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return list(UpperCamelCase_ ) def _lowerCamelCase ( self ) -> str: __lowercase : Dict = [] __lowercase : Tuple = [] __lowercase : Optional[Any] = list(self.graph )[0] stack.append(UpperCamelCase_ ) visited.append(UpperCamelCase_ ) __lowercase : Dict = -2 __lowercase : Tuple = [] __lowercase : List[Any] = s __lowercase : str = False __lowercase : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: __lowercase : Dict = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): __lowercase : Union[str, Any] = len(UpperCamelCase_ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) __lowercase : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() __lowercase : Dict = True if len(UpperCamelCase_ ) != 0: __lowercase : Optional[Any] = stack[len(UpperCamelCase_ ) - 1] else: __lowercase : Optional[int] = False indirect_parents.append(UpperCamelCase_ ) __lowercase : Optional[int] = s __lowercase : Dict = ss # check if se have reached the starting point if len(UpperCamelCase_ ) == 0: return False def _lowerCamelCase ( self ) -> Any: return list(self.graph ) def _lowerCamelCase ( self , UpperCamelCase_=-2 , UpperCamelCase_=-1 ) -> Optional[Any]: __lowercase : int = time() self.dfs(UpperCamelCase_ , UpperCamelCase_ ) __lowercase : Tuple = time() return end - begin def _lowerCamelCase ( self , UpperCamelCase_=-2 ) -> Optional[int]: __lowercase : str = time() self.bfs(UpperCamelCase_ ) __lowercase : List[Any] = time() return end - begin
249
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="audio-spectrogram-transformer" def __init__( self , UpperCamelCase_=7_68 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=30_72 , UpperCamelCase_="gelu" , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1E-12 , UpperCamelCase_=16 , UpperCamelCase_=True , UpperCamelCase_=10 , UpperCamelCase_=10 , UpperCamelCase_=10_24 , UpperCamelCase_=1_28 , **UpperCamelCase_ , ) -> Optional[int]: super().__init__(**UpperCamelCase_ ) __lowercase : Optional[Any] = hidden_size __lowercase : List[str] = num_hidden_layers __lowercase : List[str] = num_attention_heads __lowercase : Dict = intermediate_size __lowercase : List[str] = hidden_act __lowercase : Union[str, Any] = hidden_dropout_prob __lowercase : Optional[Any] = attention_probs_dropout_prob __lowercase : Dict = initializer_range __lowercase : Optional[int] = layer_norm_eps __lowercase : Optional[int] = patch_size __lowercase : List[str] = qkv_bias __lowercase : Union[str, Any] = frequency_stride __lowercase : List[Any] = time_stride __lowercase : Tuple = max_length __lowercase : int = num_mel_bins
249
1
"""simple docstring""" import numpy as np def __UpperCAmelCase ( __lowerCamelCase ) -> np.array: return 1 / (1 + np.exp(-vector )) if __name__ == "__main__": import doctest doctest.testmod()
302
"""simple docstring""" import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets lowerCAmelCase_ = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' lowerCAmelCase_ = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' lowerCAmelCase_ = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): '''simple docstring''' def UpperCAmelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,homepage='''https://github.com/krishnap25/mauve''' ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''string''' ,id='''sequence''' ), '''references''': datasets.Value('''string''' ,id='''sequence''' ), } ) ,codebase_urls=['''https://github.com/krishnap25/mauve'''] ,reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] ,) def UpperCAmelCase ( self : Dict ,_snake_case : Optional[Any] ,_snake_case : Any ,_snake_case : List[str]=None ,_snake_case : Tuple=None ,_snake_case : List[Any]=None ,_snake_case : Any=None ,_snake_case : Optional[int]="auto" ,_snake_case : Optional[int]=-1 ,_snake_case : Optional[int]=0.9 ,_snake_case : Any=5 ,_snake_case : Dict=500 ,_snake_case : Optional[int]="gpt2-large" ,_snake_case : Optional[Any]=-1 ,_snake_case : Tuple=1_024 ,_snake_case : Optional[int]=25 ,_snake_case : Dict=5 ,_snake_case : int=True ,_snake_case : Union[str, Any]=25 ,) -> Any: """simple docstring""" lowercase__ : Any = compute_mauve( p_text=_snake_case ,q_text=_snake_case ,p_features=_snake_case ,q_features=_snake_case ,p_tokens=_snake_case ,q_tokens=_snake_case ,num_buckets=_snake_case ,pca_max_data=_snake_case ,kmeans_explained_var=_snake_case ,kmeans_num_redo=_snake_case ,kmeans_max_iter=_snake_case ,featurize_model_name=_snake_case ,device_id=_snake_case ,max_text_length=_snake_case ,divergence_curve_discretization_size=_snake_case ,mauve_scaling_factor=_snake_case ,verbose=_snake_case ,seed=_snake_case ,) return out
302
1
def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' assert x is not None assert y is not None __lowerCamelCase = len(A__ ) __lowerCamelCase = len(A__ ) # declaring the array for storing the dp values __lowerCamelCase = [[0] * (n + 1) for _ in range(m + 1 )] # noqa: E741 for i in range(1 , m + 1 ): for j in range(1 , n + 1 ): __lowerCamelCase = 1 if x[i - 1] == y[j - 1] else 0 __lowerCamelCase = max(l[i - 1][j] , l[i][j - 1] , l[i - 1][j - 1] + match ) __lowerCamelCase = """""" __lowerCamelCase, __lowerCamelCase = m, n while i > 0 and j > 0: __lowerCamelCase = 1 if x[i - 1] == y[j - 1] else 0 if l[i][j] == l[i - 1][j - 1] + match: if match == 1: __lowerCamelCase = x[i - 1] + seq i -= 1 j -= 1 elif l[i][j] == l[i - 1][j]: i -= 1 else: j -= 1 return l[m][n], seq if __name__ == "__main__": UpperCAmelCase_ = 'AGGTAB' UpperCAmelCase_ = 'GXTXAYB' UpperCAmelCase_ = 4 UpperCAmelCase_ = 'GTAB' UpperCAmelCase_ , UpperCAmelCase_ = longest_common_subsequence(a, b) print('len =', ln, ', sub-sequence =', subseq) import doctest doctest.testmod()
12
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class A__(a_, unittest.TestCase ): """simple docstring""" _A : Optional[Any] = MvpTokenizer _A : List[Any] = MvpTokenizerFast _A : Dict = True _A : Optional[Any] = filter_roberta_detectors def UpperCamelCase__ ( self ) -> Union[str, Any]: super().setUp() a_ : Dict = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] a_ : int = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) a_ : List[Any] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] a_ : Optional[int] = {"""unk_token""": """<unk>"""} a_ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) a_ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowercase ) ) def UpperCamelCase__ ( self , **_lowercase ) -> Optional[Any]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def UpperCamelCase__ ( self , **_lowercase ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def UpperCamelCase__ ( self , _lowercase ) -> int: return "lower newer", "lower newer" @cached_property def UpperCamelCase__ ( self ) -> List[Any]: return MvpTokenizer.from_pretrained("""RUCAIBox/mvp""" ) @cached_property def UpperCamelCase__ ( self ) -> Union[str, Any]: return MvpTokenizerFast.from_pretrained("""RUCAIBox/mvp""" ) @require_torch def UpperCamelCase__ ( self ) -> List[str]: a_ : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] a_ : List[str] = [0, 250, 251, 17_818, 13, 39_186, 1_938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a_ : Optional[Any] = tokenizer(_lowercase , max_length=len(_lowercase ) , padding=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) a_ : Dict = batch.input_ids.tolist()[0] self.assertListEqual(_lowercase , _lowercase ) # Test that special tokens are reset @require_torch def UpperCamelCase__ ( self ) -> Union[str, Any]: a_ : List[str] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a_ : List[str] = tokenizer(_lowercase , padding=_lowercase , return_tensors="""pt""" ) # check if input_ids are returned and no labels self.assertIn("""input_ids""" , _lowercase ) self.assertIn("""attention_mask""" , _lowercase ) self.assertNotIn("""labels""" , _lowercase ) self.assertNotIn("""decoder_attention_mask""" , _lowercase ) @require_torch def UpperCamelCase__ ( self ) -> Union[str, Any]: a_ : List[Any] = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a_ : int = tokenizer(text_target=_lowercase , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def UpperCamelCase__ ( self ) -> Any: for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a_ : int = tokenizer( ["""I am a small frog""" * 1_024, """I am a small frog"""] , padding=_lowercase , truncation=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(batch.input_ids.shape , (2, 1_024) ) @require_torch def UpperCamelCase__ ( self ) -> List[str]: a_ : Tuple = ["""A long paragraph for summarization."""] a_ : Optional[Any] = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a_ : List[Any] = tokenizer(_lowercase , text_target=_lowercase , return_tensors="""pt""" ) a_ : Union[str, Any] = inputs["""input_ids"""] a_ : Dict = inputs["""labels"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def UpperCamelCase__ ( self ) -> int: pass def UpperCamelCase__ ( self ) -> Union[str, Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): a_ : List[str] = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) a_ : List[str] = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) a_ : Optional[int] = """A, <mask> AllenNLP sentence.""" a_ : Union[str, Any] = tokenizer_r.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) a_ : Tuple = tokenizer_p.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) a_ : List[Any] = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) a_ : str = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
248
0
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( snake_case : str , snake_case : str ) -> float: def get_matched_characters(snake_case : str , snake_case : str ) -> str: a : Optional[int] = [] a : Dict = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): a : Tuple = int(max(0 , i - limit ) ) a : List[str] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(_UpperCamelCase ) a : List[Any] = F"""{_stra[0:_stra.index(_UpperCamelCase )]} {_stra[_stra.index(_UpperCamelCase ) + 1:]}""" return "".join(_UpperCamelCase ) # matching characters a : Tuple = get_matched_characters(_UpperCamelCase , _UpperCamelCase ) a : Dict = get_matched_characters(_UpperCamelCase , _UpperCamelCase ) a : List[str] = len(_UpperCamelCase ) # transposition a : Optional[int] = ( len([(ca, ca) for ca, ca in zip(_UpperCamelCase , _UpperCamelCase ) if ca != ca] ) // 2 ) if not match_count: a : Optional[int] = 0.0 else: a : Optional[int] = ( 1 / 3 * ( match_count / len(_UpperCamelCase ) + match_count / len(_UpperCamelCase ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters a : Tuple = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("""hello""", """world"""))
359
'''simple docstring''' from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def SCREAMING_SNAKE_CASE__ ( snake_case : str , snake_case : float | Decimal , snake_case : float = 10**-10 ) -> float: """simple docstring""" a : Dict = a while True: a : Any = Decimal(snake_case ) - ( Decimal(eval(snake_case ) ) / Decimal(eval(str(diff(snake_case ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(snake_case ) ) < precision: # noqa: S307 return float(snake_case ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(f'''The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}''') # Find root of polynomial print(f'''The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}''') # Find Square Root of 5 print(f'''The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}''') # Exponential Roots print(f'''The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}''')
345
0
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : int = [] __magic_name__ : List[Any] = [] __magic_name__ : Optional[Any] = [] for rt in rc.restypes: __magic_name__ : str = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __magic_name__ : str = {name: i for i, name in enumerate(_A )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) __magic_name__ : List[str] = torch.tensor( _A, dtype=torch.intaa, device=protein["""aatype"""].device, ) __magic_name__ : Dict = torch.tensor( _A, dtype=torch.intaa, device=protein["""aatype"""].device, ) __magic_name__ : Tuple = torch.tensor( _A, dtype=torch.floataa, device=protein["""aatype"""].device, ) __magic_name__ : Dict = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __magic_name__ : str = restype_atomaa_to_atomaa[protein_aatype] __magic_name__ : Union[str, Any] = restype_atomaa_mask[protein_aatype] __magic_name__ : List[Any] = residx_atomaa_mask __magic_name__ : str = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __magic_name__ : int = restype_atomaa_to_atomaa[protein_aatype] __magic_name__ : Optional[Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __magic_name__ : List[Any] = torch.zeros([21, 37], dtype=torch.floataa, device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): __magic_name__ : str = rc.restype_atoa[restype_letter] __magic_name__ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: __magic_name__ : Dict = rc.atom_order[atom_name] __magic_name__ : List[str] = 1 __magic_name__ : List[Any] = restype_atomaa_mask[protein_aatype] __magic_name__ : Optional[Any] = residx_atomaa_mask return protein def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : List[Any] = tree_map(lambda _A : torch.tensor(_A, device=batch["""aatype"""].device ), _A, np.ndarray ) __magic_name__ : Optional[Any] = tensor_tree_map(lambda _A : np.array(_A ), make_atomaa_masks(_A ) ) return out
342
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase ( _A, _A, _A ): """simple docstring""" __magic_name__ : Dict = MobileBertConfig.from_json_file(_A ) print(f'Building PyTorch model from configuration: {config}' ) __magic_name__ : Tuple = MobileBertForPreTraining(_A ) # Load weights from tf checkpoint __magic_name__ : int = load_tf_weights_in_mobilebert(_A, _A, _A ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict(), _A ) if __name__ == "__main__": __magic_name__: Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __magic_name__: Dict = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
342
1
import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class __a( snake_case__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase = FlaxAutoencoderKL @property def a__ ( self ) -> Optional[int]: UpperCAmelCase_ : List[Any] = 4 UpperCAmelCase_ : Optional[Any] = 3 UpperCAmelCase_ : Optional[Any] = (32, 32) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Any = jax.random.uniform(_A ,((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def a__ ( self ) -> List[Any]: UpperCAmelCase_ : Any = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } UpperCAmelCase_ : int = self.dummy_input return init_dict, inputs_dict
364
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: __a = None __a = logging.get_logger(__name__) __a = '▁' __a = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} __a = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'}, 'tokenizer_file': { 'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json' }, } __a = { 'google/pegasus-xsum': 512, } class __a( _a ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = PegasusTokenizer lowerCAmelCase = ['''input_ids''', '''attention_mask'''] def __init__( self ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE="<pad>" ,_SCREAMING_SNAKE_CASE="</s>" ,_SCREAMING_SNAKE_CASE="<unk>" ,_SCREAMING_SNAKE_CASE="<mask_2>" ,_SCREAMING_SNAKE_CASE="<mask_1>" ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE=103 ,**_SCREAMING_SNAKE_CASE ,) -> Optional[Any]: UpperCAmelCase_ : Dict = offset if additional_special_tokens is not None: if not isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): raise TypeError( f'''additional_special_tokens should be of type {type(_SCREAMING_SNAKE_CASE )}, but is''' f''' {type(_SCREAMING_SNAKE_CASE )}''' ) UpperCAmelCase_ : str = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'''<unk_{i}>''' for i in range(len(_SCREAMING_SNAKE_CASE ) ,self.offset - 1 ) ] if len(set(_SCREAMING_SNAKE_CASE ) ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) UpperCAmelCase_ : int = additional_special_tokens_extended else: UpperCAmelCase_ : Any = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'''<unk_{i}>''' for i in range(2 ,self.offset )] super().__init__( _SCREAMING_SNAKE_CASE ,tokenizer_file=_SCREAMING_SNAKE_CASE ,pad_token=_SCREAMING_SNAKE_CASE ,eos_token=_SCREAMING_SNAKE_CASE ,unk_token=_SCREAMING_SNAKE_CASE ,mask_token=_SCREAMING_SNAKE_CASE ,mask_token_sent=_SCREAMING_SNAKE_CASE ,offset=_SCREAMING_SNAKE_CASE ,additional_special_tokens=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ,) UpperCAmelCase_ : str = vocab_file UpperCAmelCase_ : Dict = False if not self.vocab_file else True def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> Any: UpperCAmelCase_ : Dict = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( '''There should be 3 special tokens: mask_token, pad_token, and eos_token +''' f''' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}''' ) return [1 if x in all_special_ids else 0 for x in seq] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = False ) -> List[int]: if already_has_special_tokens: return self._special_token_mask(_SCREAMING_SNAKE_CASE ) elif token_ids_a is None: return self._special_token_mask(_SCREAMING_SNAKE_CASE ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : Optional[int] = os.path.join( _SCREAMING_SNAKE_CASE ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file ,_SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
235
0
'''simple docstring''' from importlib import import_module from .logging import get_logger __lowerCAmelCase = get_logger(__name__) class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None ) -> int: _snake_case = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) _snake_case = module._original_module if isinstance(lowerCAmelCase__ , _PatchedModuleObj ) else module class _lowerCAmelCase : '''simple docstring''' lowerCAmelCase_ = [] def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None ) -> List[Any]: _snake_case = obj _snake_case = target _snake_case = new _snake_case = target.split(""".""" )[0] _snake_case = {} _snake_case = attrs or [] def __enter__(self ) -> Dict: _snake_case = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowerCAmelCase__ ) ): try: _snake_case = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _snake_case = getattr(self.obj , lowerCAmelCase__ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowerCAmelCase__ , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): _snake_case = obj_attr # patch at top level setattr(self.obj , lowerCAmelCase__ , _PatchedModuleObj(lowerCAmelCase__ , attrs=self.attrs ) ) _snake_case = getattr(self.obj , lowerCAmelCase__ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowerCAmelCase__ , lowerCAmelCase__ , _PatchedModuleObj(getattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , attrs=self.attrs ) ) _snake_case = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) # finally set the target attribute setattr(lowerCAmelCase__ , lowerCAmelCase__ , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _snake_case = getattr(import_module(""".""".join(lowerCAmelCase__ ) ) , lowerCAmelCase__ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , lowerCAmelCase__ ) is attr_value: _snake_case = getattr(self.obj , lowerCAmelCase__ ) setattr(self.obj , lowerCAmelCase__ , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _snake_case = globals()['''__builtins__'''][target_attr] setattr(self.obj , lowerCAmelCase__ , self.new ) else: raise RuntimeError(f"""Tried to patch attribute {target_attr} instead of a submodule.""" ) def __exit__(self , *UpperCAmelCase ) -> Optional[int]: for attr in list(self.original ): setattr(self.obj , lowerCAmelCase__ , self.original.pop(lowerCAmelCase__ ) ) def lowercase (self ) -> List[Any]: self.__enter__() self._active_patches.append(self ) def lowercase (self ) -> int: try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
341
from __future__ import annotations import bisect def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ = 0 , lowercase__ = -1 ): if hi < 0: __SCREAMING_SNAKE_CASE : Union[str, Any] = len(lowercase__ ) while lo < hi: __SCREAMING_SNAKE_CASE : Any = lo + (hi - lo) // 2 if sorted_collection[mid] < item: __SCREAMING_SNAKE_CASE : Union[str, Any] = mid + 1 else: __SCREAMING_SNAKE_CASE : Optional[Any] = mid return lo def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ = 0 , lowercase__ = -1 ): if hi < 0: __SCREAMING_SNAKE_CASE : List[Any] = len(lowercase__ ) while lo < hi: __SCREAMING_SNAKE_CASE : Optional[int] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: __SCREAMING_SNAKE_CASE : Any = mid + 1 else: __SCREAMING_SNAKE_CASE : Optional[int] = mid return lo def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ = 0 , lowercase__ = -1 ): sorted_collection.insert(bisect_left(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) , lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ = 0 , lowercase__ = -1 ): sorted_collection.insert(bisect_right(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) , lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Any = 0 __SCREAMING_SNAKE_CASE : List[Any] = len(lowercase__ ) - 1 while left <= right: __SCREAMING_SNAKE_CASE : str = left + (right - left) // 2 __SCREAMING_SNAKE_CASE : List[str] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: __SCREAMING_SNAKE_CASE : int = midpoint - 1 else: __SCREAMING_SNAKE_CASE : Dict = midpoint + 1 return None def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = bisect.bisect_left(lowercase__ , lowercase__ ) if index != len(lowercase__ ) and sorted_collection[index] == item: return index return None def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if right < left: return None __SCREAMING_SNAKE_CASE : int = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowercase__ , lowercase__ , lowercase__ , midpoint - 1 ) else: return binary_search_by_recursion(lowercase__ , lowercase__ , midpoint + 1 , lowercase__ ) if __name__ == "__main__": __lowerCAmelCase : Dict =input('Enter numbers separated by comma:\n').strip() __lowerCAmelCase : str =sorted(int(item) for item in user_input.split(',')) __lowerCAmelCase : Tuple =int(input('Enter a single number to be found in the list:\n')) __lowerCAmelCase : Tuple =binary_search(collection, target) if result is None: print(f"""{target} was not found in {collection}.""") else: print(f"""{target} was found at position {result} in {collection}.""")
9
0
import os import posixpath import uuid from dataclasses import dataclass from typing import TYPE_CHECKING, Iterable, List, Optional, Tuple, Union import numpy as np import pyarrow as pa import datasets from datasets.arrow_writer import ArrowWriter, ParquetWriter from datasets.config import MAX_SHARD_SIZE from datasets.filesystems import ( is_remote_filesystem, rename, ) from datasets.iterable_dataset import _BaseExamplesIterable from datasets.utils.py_utils import convert_file_size_to_int _lowerCamelCase : int = datasets.utils.logging.get_logger(__name__) if TYPE_CHECKING: import pyspark @dataclass class lowerCamelCase (datasets.BuilderConfig ): """simple docstring""" UpperCAmelCase_ = None def _a ( SCREAMING_SNAKE_CASE__ : "pyspark.sql.DataFrame" , SCREAMING_SNAKE_CASE__ : List[int] , ) -> Optional[Any]: '''simple docstring''' import pyspark def generate_fn(): SCREAMING_SNAKE_CASE__ : Optional[int] = df.select("*" , pyspark.sql.functions.spark_partition_id().alias("part_id" ) ) for partition_id in partition_order: SCREAMING_SNAKE_CASE__ : List[str] = df_with_partition_id.select("*" ).where(f'''part_id = {partition_id}''' ).drop("part_id" ) SCREAMING_SNAKE_CASE__ : Dict = partition_df.collect() SCREAMING_SNAKE_CASE__ : List[str] = 0 for row in rows: yield f'''{partition_id}_{row_id}''', row.asDict() row_id += 1 return generate_fn class lowerCamelCase (_BaseExamplesIterable ): """simple docstring""" def __init__( self : Tuple, _UpperCAmelCase : "pyspark.sql.DataFrame", _UpperCAmelCase : List[Any]=None, ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = df SCREAMING_SNAKE_CASE__ : Tuple = partition_order or range(self.df.rdd.getNumPartitions() ) SCREAMING_SNAKE_CASE__ : str = _generate_iterable_examples(self.df, self.partition_order ) def __iter__( self : Dict ) -> Dict: """simple docstring""" yield from self.generate_examples_fn() def A_ ( self : List[Any], _UpperCAmelCase : np.random.Generator ) -> "SparkExamplesIterable": """simple docstring""" SCREAMING_SNAKE_CASE__ : str = list(range(self.df.rdd.getNumPartitions() ) ) generator.shuffle(_UpperCAmelCase ) return SparkExamplesIterable(self.df, partition_order=_UpperCAmelCase ) def A_ ( self : Any, _UpperCAmelCase : int, _UpperCAmelCase : int ) -> "SparkExamplesIterable": """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = self.split_shard_indices_by_worker(_UpperCAmelCase, _UpperCAmelCase ) return SparkExamplesIterable(self.df, partition_order=_UpperCAmelCase ) @property def A_ ( self : Any ) -> int: """simple docstring""" return len(self.partition_order ) class lowerCamelCase (datasets.DatasetBuilder ): """simple docstring""" UpperCAmelCase_ = SparkConfig def __init__( self : Dict, _UpperCAmelCase : "pyspark.sql.DataFrame", _UpperCAmelCase : str = None, _UpperCAmelCase : str = None, **_UpperCAmelCase : Any, ) -> Optional[Any]: """simple docstring""" import pyspark SCREAMING_SNAKE_CASE__ : str = pyspark.sql.SparkSession.builder.getOrCreate() SCREAMING_SNAKE_CASE__ : int = df SCREAMING_SNAKE_CASE__ : Dict = working_dir super().__init__( cache_dir=_UpperCAmelCase, config_name=str(self.df.semanticHash() ), **_UpperCAmelCase, ) def A_ ( self : List[Any] ) -> List[Any]: """simple docstring""" # Returns the path of the created file. def create_cache_and_write_probe(_UpperCAmelCase : Tuple ): # makedirs with exist_ok will recursively create the directory. It will not throw an error if directories # already exist. os.makedirs(self._cache_dir, exist_ok=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = os.path.join(self._cache_dir, "fs_test" + uuid.uuida().hex ) # Opening the file in append mode will create a new file unless it already exists, in which case it will not # change the file contents. open(_UpperCAmelCase, "a" ) return [probe_file] if self._spark.conf.get("spark.master", "" ).startswith("local" ): return # If the cluster is multi-node, make sure that the user provided a cache_dir and that it is on an NFS # accessible to the driver. # TODO: Stream batches to the driver using ArrowCollectSerializer instead of throwing an error. if self._cache_dir: SCREAMING_SNAKE_CASE__ : Tuple = ( self._spark.sparkContext.parallelize(range(1 ), 1 ).mapPartitions(_UpperCAmelCase ).collect() ) if os.path.isfile(probe[0] ): return raise ValueError( "When using Dataset.from_spark on a multi-node cluster, the driver and all workers should be able to access cache_dir" ) def A_ ( self : str ) -> str: """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def A_ ( self : List[str], _UpperCAmelCase : datasets.download.download_manager.DownloadManager ) -> int: """simple docstring""" return [datasets.SplitGenerator(name=datasets.Split.TRAIN )] def A_ ( self : Dict, _UpperCAmelCase : str ) -> Optional[Any]: """simple docstring""" import pyspark def get_arrow_batch_size(_UpperCAmelCase : Any ): for batch in it: yield pa.RecordBatch.from_pydict({"batch_bytes": [batch.nbytes]} ) SCREAMING_SNAKE_CASE__ : List[Any] = self.df.count() SCREAMING_SNAKE_CASE__ : int = df_num_rows if df_num_rows <= 1_0_0 else 1_0_0 # Approximate the size of each row (in Arrow format) by averaging over a max-100-row sample. SCREAMING_SNAKE_CASE__ : Union[str, Any] = ( self.df.limit(_UpperCAmelCase ) .repartition(1 ) .mapInArrow(_UpperCAmelCase, "batch_bytes: long" ) .agg(pyspark.sql.functions.sum("batch_bytes" ).alias("sample_bytes" ) ) .collect()[0] .sample_bytes / sample_num_rows ) SCREAMING_SNAKE_CASE__ : List[Any] = approx_bytes_per_row * df_num_rows if approx_total_size > max_shard_size: # Make sure there is at least one row per partition. SCREAMING_SNAKE_CASE__ : Any = min(_UpperCAmelCase, int(approx_total_size / max_shard_size ) ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.df.repartition(_UpperCAmelCase ) def A_ ( self : Union[str, Any], _UpperCAmelCase : str, _UpperCAmelCase : str, _UpperCAmelCase : int, ) -> Iterable[Tuple[int, bool, Union[int, tuple]]]: """simple docstring""" import pyspark SCREAMING_SNAKE_CASE__ : int = ParquetWriter if file_format == "parquet" else ArrowWriter SCREAMING_SNAKE_CASE__ : str = os.path.join(self._working_dir, os.path.basename(_UpperCAmelCase ) ) if self._working_dir else fpath SCREAMING_SNAKE_CASE__ : Tuple = file_format == "parquet" # Define these so that we don't reference self in write_arrow, which will result in a pickling error due to # pickling the SparkContext. SCREAMING_SNAKE_CASE__ : Optional[int] = self.config.features SCREAMING_SNAKE_CASE__ : Dict = self._writer_batch_size SCREAMING_SNAKE_CASE__ : int = self._fs.storage_options def write_arrow(_UpperCAmelCase : Union[str, Any] ): # Within the same SparkContext, no two task attempts will share the same attempt ID. SCREAMING_SNAKE_CASE__ : Any = pyspark.TaskContext().taskAttemptId() SCREAMING_SNAKE_CASE__ : Union[str, Any] = next(_UpperCAmelCase, _UpperCAmelCase ) if first_batch is None: # Some partitions might not receive any data. return pa.RecordBatch.from_arrays( [[task_id], [0], [0]], names=["task_id", "num_examples", "num_bytes"], ) SCREAMING_SNAKE_CASE__ : int = 0 SCREAMING_SNAKE_CASE__ : Tuple = writer_class( features=_UpperCAmelCase, path=working_fpath.replace("SSSSS", F'''{shard_id:05d}''' ).replace("TTTTT", F'''{task_id:05d}''' ), writer_batch_size=_UpperCAmelCase, storage_options=_UpperCAmelCase, embed_local_files=_UpperCAmelCase, ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = pa.Table.from_batches([first_batch] ) writer.write_table(_UpperCAmelCase ) for batch in it: if max_shard_size is not None and writer._num_bytes >= max_shard_size: SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : Any = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]], names=["task_id", "num_examples", "num_bytes"], ) shard_id += 1 SCREAMING_SNAKE_CASE__ : Optional[int] = writer_class( features=writer._features, path=working_fpath.replace("SSSSS", F'''{shard_id:05d}''' ).replace("TTTTT", F'''{task_id:05d}''' ), writer_batch_size=_UpperCAmelCase, storage_options=_UpperCAmelCase, embed_local_files=_UpperCAmelCase, ) SCREAMING_SNAKE_CASE__ : str = pa.Table.from_batches([batch] ) writer.write_table(_UpperCAmelCase ) if writer._num_bytes > 0: SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : str = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]], names=["task_id", "num_examples", "num_bytes"], ) if working_fpath != fpath: for file in os.listdir(os.path.dirname(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE__ : Dict = os.path.join(os.path.dirname(_UpperCAmelCase ), os.path.basename(_UpperCAmelCase ) ) shutil.move(_UpperCAmelCase, _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Any = ( self.df.mapInArrow(_UpperCAmelCase, "task_id: long, num_examples: long, num_bytes: long" ) .groupBy("task_id" ) .agg( pyspark.sql.functions.sum("num_examples" ).alias("total_num_examples" ), pyspark.sql.functions.sum("num_bytes" ).alias("total_num_bytes" ), pyspark.sql.functions.count("num_bytes" ).alias("num_shards" ), pyspark.sql.functions.collect_list("num_examples" ).alias("shard_lengths" ), ) .collect() ) for row in stats: yield row.task_id, (row.total_num_examples, row.total_num_bytes, row.num_shards, row.shard_lengths) def A_ ( self : List[str], _UpperCAmelCase : "datasets.SplitGenerator", _UpperCAmelCase : str = "arrow", _UpperCAmelCase : Optional[Union[str, int]] = None, _UpperCAmelCase : Optional[int] = None, **_UpperCAmelCase : Any, ) -> int: """simple docstring""" self._validate_cache_dir() SCREAMING_SNAKE_CASE__ : Dict = convert_file_size_to_int(max_shard_size or MAX_SHARD_SIZE ) self._repartition_df_if_needed(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Any = not is_remote_filesystem(self._fs ) SCREAMING_SNAKE_CASE__ : Tuple = os.path.join if is_local else posixpath.join SCREAMING_SNAKE_CASE__ : int = "-TTTTT-SSSSS-of-NNNNN" SCREAMING_SNAKE_CASE__ : Tuple = F'''{self.name}-{split_generator.name}{SUFFIX}.{file_format}''' SCREAMING_SNAKE_CASE__ : Dict = path_join(self._output_dir, _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Dict = 0 SCREAMING_SNAKE_CASE__ : str = 0 SCREAMING_SNAKE_CASE__ : Union[str, Any] = 0 SCREAMING_SNAKE_CASE__ : Any = [] SCREAMING_SNAKE_CASE__ : str = [] for task_id, content in self._prepare_split_single(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ): ( ( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) ,( SCREAMING_SNAKE_CASE__ ) , ) : Tuple = content if num_bytes > 0: total_num_examples += num_examples total_num_bytes += num_bytes total_shards += num_shards task_id_and_num_shards.append((task_id, num_shards) ) all_shard_lengths.extend(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Dict = total_num_examples SCREAMING_SNAKE_CASE__ : List[str] = total_num_bytes # should rename everything at the end logger.debug(F'''Renaming {total_shards} shards.''' ) if total_shards > 1: SCREAMING_SNAKE_CASE__ : Dict = all_shard_lengths # Define fs outside of _rename_shard so that we don't reference self in the function, which will result in a # pickling error due to pickling the SparkContext. SCREAMING_SNAKE_CASE__ : str = self._fs # use the -SSSSS-of-NNNNN pattern def _rename_shard( _UpperCAmelCase : int, _UpperCAmelCase : int, _UpperCAmelCase : int, ): rename( _UpperCAmelCase, fpath.replace("SSSSS", F'''{shard_id:05d}''' ).replace("TTTTT", F'''{task_id:05d}''' ), fpath.replace("TTTTT-SSSSS", F'''{global_shard_id:05d}''' ).replace("NNNNN", F'''{total_shards:05d}''' ), ) SCREAMING_SNAKE_CASE__ : List[str] = [] SCREAMING_SNAKE_CASE__ : List[Any] = 0 for i in range(len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : Dict = task_id_and_num_shards[i] for shard_id in range(_UpperCAmelCase ): args.append([task_id, shard_id, global_shard_id] ) global_shard_id += 1 self._spark.sparkContext.parallelize(_UpperCAmelCase, len(_UpperCAmelCase ) ).map(lambda _UpperCAmelCase : _rename_shard(*_UpperCAmelCase ) ).collect() else: # don't use any pattern SCREAMING_SNAKE_CASE__ : List[str] = 0 SCREAMING_SNAKE_CASE__ : List[Any] = task_id_and_num_shards[0][0] self._rename( fpath.replace("SSSSS", F'''{shard_id:05d}''' ).replace("TTTTT", F'''{task_id:05d}''' ), fpath.replace(_UpperCAmelCase, "" ), ) def A_ ( self : Any, _UpperCAmelCase : "datasets.SplitGenerator", ) -> SparkExamplesIterable: """simple docstring""" return SparkExamplesIterable(self.df )
191
def _a ( SCREAMING_SNAKE_CASE__ : str ) -> str: '''simple docstring''' if not all(char in "01" for char in bin_string ): raise ValueError("Non-binary value was passed to the function" ) if not bin_string: raise ValueError("Empty string was passed to the function" ) SCREAMING_SNAKE_CASE__ : List[Any] = "" while len(SCREAMING_SNAKE_CASE__ ) % 3 != 0: SCREAMING_SNAKE_CASE__ : str = "0" + bin_string SCREAMING_SNAKE_CASE__ : List[Any] = [ bin_string[index : index + 3] for index in range(len(SCREAMING_SNAKE_CASE__ ) ) if index % 3 == 0 ] for bin_group in bin_string_in_3_list: SCREAMING_SNAKE_CASE__ : List[Any] = 0 for index, val in enumerate(SCREAMING_SNAKE_CASE__ ): oct_val += int(2 ** (2 - index) * int(SCREAMING_SNAKE_CASE__ ) ) oct_string += str(SCREAMING_SNAKE_CASE__ ) return oct_string if __name__ == "__main__": from doctest import testmod testmod()
191
1
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' def UpperCamelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """width_multiplier""" ) ) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any , lowerCAmelCase__ : int , lowerCAmelCase__ : List[str]=1_3 , lowerCAmelCase__ : List[Any]=6_4 , lowerCAmelCase__ : List[Any]=2 , lowerCAmelCase__ : Dict=3 , lowerCAmelCase__ : Any="swish" , lowerCAmelCase__ : List[str]=3 , lowerCAmelCase__ : List[Any]=3_2 , lowerCAmelCase__ : Dict=0.1 , lowerCAmelCase__ : Optional[int]=0.02 , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : Any=1_0 , lowerCAmelCase__ : Optional[int]=None , lowerCAmelCase__ : Optional[Any]=0.25 , lowerCAmelCase__ : Optional[int]=0.0 , lowerCAmelCase__ : Dict=0.0 , ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = parent __SCREAMING_SNAKE_CASE : Tuple = batch_size __SCREAMING_SNAKE_CASE : Tuple = image_size __SCREAMING_SNAKE_CASE : List[Any] = patch_size __SCREAMING_SNAKE_CASE : Union[str, Any] = num_channels __SCREAMING_SNAKE_CASE : Tuple = make_divisible(5_1_2 * width_multiplier , divisor=8 ) __SCREAMING_SNAKE_CASE : int = hidden_act __SCREAMING_SNAKE_CASE : List[str] = conv_kernel_size __SCREAMING_SNAKE_CASE : str = output_stride __SCREAMING_SNAKE_CASE : Tuple = classifier_dropout_prob __SCREAMING_SNAKE_CASE : Dict = use_labels __SCREAMING_SNAKE_CASE : Optional[int] = is_training __SCREAMING_SNAKE_CASE : List[Any] = num_labels __SCREAMING_SNAKE_CASE : int = initializer_range __SCREAMING_SNAKE_CASE : Any = scope __SCREAMING_SNAKE_CASE : List[Any] = width_multiplier __SCREAMING_SNAKE_CASE : Any = ffn_dropout __SCREAMING_SNAKE_CASE : Optional[Any] = attn_dropout def UpperCamelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __SCREAMING_SNAKE_CASE : Any = None __SCREAMING_SNAKE_CASE : Any = None if self.use_labels: __SCREAMING_SNAKE_CASE : Any = ids_tensor([self.batch_size] , self.num_labels ) __SCREAMING_SNAKE_CASE : List[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __SCREAMING_SNAKE_CASE : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def UpperCamelCase__ ( self : Union[str, Any] ): """simple docstring""" return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def UpperCamelCase__ ( self : int , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = MobileViTVaModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE : str = model(lowerCAmelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCamelCase__ ( self : List[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = self.num_labels __SCREAMING_SNAKE_CASE : Union[str, Any] = MobileViTVaForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE : int = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self : Union[str, Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = self.num_labels __SCREAMING_SNAKE_CASE : Optional[Any] = MobileViTVaForSemanticSegmentation(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE : List[str] = model(lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __SCREAMING_SNAKE_CASE : Optional[int] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCamelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = config_and_inputs __SCREAMING_SNAKE_CASE : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _UpperCamelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' _A : Dict = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) _A : str = ( { '''feature-extraction''': MobileViTVaModel, '''image-classification''': MobileViTVaForImageClassification, '''image-segmentation''': MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) _A : Optional[Any] = False _A : Tuple = False _A : str = False _A : Union[str, Any] = False def UpperCamelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = MobileViTVaModelTester(self ) __SCREAMING_SNAKE_CASE : List[Any] = MobileViTVaConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def UpperCamelCase__ ( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""MobileViTV2 does not use inputs_embeds""" ) def UpperCamelCase__ ( self : List[Any] ): """simple docstring""" pass @unittest.skip(reason="""MobileViTV2 does not support input and output embeddings""" ) def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" pass @unittest.skip(reason="""MobileViTV2 does not output attentions""" ) def UpperCamelCase__ ( self : Any ): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason="""Got `CUDA error: misaligned address` for tests after this one being run.""" ) def UpperCamelCase__ ( self : Tuple ): """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase__ ( self : List[str] ): """simple docstring""" pass def UpperCamelCase__ ( self : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE : Dict = model_class(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE : List[Any] = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def UpperCamelCase__ ( self : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCamelCase__ ( self : Dict ): """simple docstring""" def check_hidden_states_output(lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Dict , lowerCAmelCase__ : List[Any] ): __SCREAMING_SNAKE_CASE : Optional[int] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE : List[str] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) __SCREAMING_SNAKE_CASE : List[Any] = outputs.hidden_states __SCREAMING_SNAKE_CASE : Union[str, Any] = 5 self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __SCREAMING_SNAKE_CASE : int = 2 for i in range(len(lowerCAmelCase__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE : str = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) def UpperCamelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase__ ) @slow def UpperCamelCase__ ( self : str ): """simple docstring""" for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE : Tuple = MobileViTVaModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def lowerCAmelCase_ ( ): __SCREAMING_SNAKE_CASE : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self : int ): """simple docstring""" return ( MobileViTImageProcessor.from_pretrained("""apple/mobilevitv2-1.0-imagenet1k-256""" ) if is_vision_available() else None ) @slow def UpperCamelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = MobileViTVaForImageClassification.from_pretrained("""apple/mobilevitv2-1.0-imagenet1k-256""" ).to( lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = self.default_image_processor __SCREAMING_SNAKE_CASE : int = prepare_img() __SCREAMING_SNAKE_CASE : Any = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE : str = model(**lowerCAmelCase__ ) # verify the logits __SCREAMING_SNAKE_CASE : Optional[Any] = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = torch.tensor([-1.63_36E00, -7.32_04E-02, -5.18_83E-01] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def UpperCamelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = MobileViTVaForSemanticSegmentation.from_pretrained("""shehan97/mobilevitv2-1.0-voc-deeplabv3""" ) __SCREAMING_SNAKE_CASE : Any = model.to(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = MobileViTImageProcessor.from_pretrained("""shehan97/mobilevitv2-1.0-voc-deeplabv3""" ) __SCREAMING_SNAKE_CASE : Dict = prepare_img() __SCREAMING_SNAKE_CASE : int = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE : Dict = model(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = outputs.logits # verify the logits __SCREAMING_SNAKE_CASE : int = torch.Size((1, 2_1, 3_2, 3_2) ) self.assertEqual(logits.shape , lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = torch.tensor( [ [[7.08_63, 7.15_25, 6.82_01], [6.69_31, 6.87_70, 6.89_33], [6.29_78, 7.03_66, 6.96_36]], [[-3.71_34, -3.67_12, -3.66_75], [-3.58_25, -3.35_49, -3.47_77], [-3.34_35, -3.39_79, -3.28_57]], [[-2.93_29, -2.80_03, -2.73_69], [-3.05_64, -2.47_80, -2.02_07], [-2.68_89, -1.92_98, -1.76_40]], ] , device=lowerCAmelCase__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def UpperCamelCase__ ( self : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = MobileViTVaForSemanticSegmentation.from_pretrained("""shehan97/mobilevitv2-1.0-voc-deeplabv3""" ) __SCREAMING_SNAKE_CASE : List[Any] = model.to(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = MobileViTImageProcessor.from_pretrained("""shehan97/mobilevitv2-1.0-voc-deeplabv3""" ) __SCREAMING_SNAKE_CASE : Optional[int] = prepare_img() __SCREAMING_SNAKE_CASE : List[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE : Union[str, Any] = model(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = outputs.logits.detach().cpu() __SCREAMING_SNAKE_CASE : str = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__ , target_sizes=[(5_0, 6_0)] ) __SCREAMING_SNAKE_CASE : int = torch.Size((5_0, 6_0) ) self.assertEqual(segmentation[0].shape , lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = torch.Size((3_2, 3_2) ) self.assertEqual(segmentation[0].shape , lowerCAmelCase__ )
112
'''simple docstring''' import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging UpperCamelCase__ : List[str] = logging.get_logger(__name__) UpperCamelCase__ : Tuple = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED UpperCamelCase__ : Tuple = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } UpperCamelCase__ : List[Any] = { '''allenai/led-base-16384''': 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowerCAmelCase_ ( ): __SCREAMING_SNAKE_CASE : Tuple = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) __SCREAMING_SNAKE_CASE : Any = bs[:] __SCREAMING_SNAKE_CASE : Tuple = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCamelCase ) cs.append(2**8 + n ) n += 1 __SCREAMING_SNAKE_CASE : Union[str, Any] = [chr(_lowerCamelCase ) for n in cs] return dict(zip(_lowerCamelCase , _lowerCamelCase ) ) def lowerCAmelCase_ ( _lowerCamelCase: Optional[int] ): __SCREAMING_SNAKE_CASE : Dict = set() __SCREAMING_SNAKE_CASE : Optional[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __SCREAMING_SNAKE_CASE : str = char return pairs class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' _A : Union[str, Any] = VOCAB_FILES_NAMES _A : Any = PRETRAINED_VOCAB_FILES_MAP _A : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : Optional[Any] = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Union[str, Any]="replace" , lowerCAmelCase__ : Dict="<s>" , lowerCAmelCase__ : List[str]="</s>" , lowerCAmelCase__ : Tuple="</s>" , lowerCAmelCase__ : Tuple="<s>" , lowerCAmelCase__ : Union[str, Any]="<unk>" , lowerCAmelCase__ : Union[str, Any]="<pad>" , lowerCAmelCase__ : int="<mask>" , lowerCAmelCase__ : str=False , **lowerCAmelCase__ : int , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else bos_token __SCREAMING_SNAKE_CASE : List[str] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else eos_token __SCREAMING_SNAKE_CASE : List[str] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else sep_token __SCREAMING_SNAKE_CASE : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else cls_token __SCREAMING_SNAKE_CASE : Optional[Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else unk_token __SCREAMING_SNAKE_CASE : Optional[int] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it __SCREAMING_SNAKE_CASE : Optional[int] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ , **lowerCAmelCase__ , ) with open(lowerCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: __SCREAMING_SNAKE_CASE : str = json.load(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = {v: k for k, v in self.encoder.items()} __SCREAMING_SNAKE_CASE : Dict = errors # how to handle errors in decoding __SCREAMING_SNAKE_CASE : Union[str, Any] = bytes_to_unicode() __SCREAMING_SNAKE_CASE : Tuple = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ , encoding="""utf-8""" ) as merges_handle: __SCREAMING_SNAKE_CASE : Optional[Any] = merges_handle.read().split("""\n""" )[1:-1] __SCREAMING_SNAKE_CASE : int = [tuple(merge.split() ) for merge in bpe_merges] __SCREAMING_SNAKE_CASE : Optional[int] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) __SCREAMING_SNAKE_CASE : int = {} __SCREAMING_SNAKE_CASE : Any = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions __SCREAMING_SNAKE_CASE : str = re.compile(r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCamelCase__ ( self : Union[str, Any] ): """simple docstring""" return len(self.encoder ) def UpperCamelCase__ ( self : Any ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def UpperCamelCase__ ( self : Dict , lowerCAmelCase__ : Any ): """simple docstring""" if token in self.cache: return self.cache[token] __SCREAMING_SNAKE_CASE : Union[str, Any] = tuple(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Dict = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: __SCREAMING_SNAKE_CASE : Union[str, Any] = min(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = bigram __SCREAMING_SNAKE_CASE : List[Any] = [] __SCREAMING_SNAKE_CASE : Optional[int] = 0 while i < len(lowerCAmelCase__ ): try: __SCREAMING_SNAKE_CASE : Dict = word.index(lowerCAmelCase__ , lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __SCREAMING_SNAKE_CASE : Dict = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __SCREAMING_SNAKE_CASE : Tuple = tuple(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = new_word if len(lowerCAmelCase__ ) == 1: break else: __SCREAMING_SNAKE_CASE : Union[str, Any] = get_pairs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = """ """.join(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = word return word def UpperCamelCase__ ( self : Optional[Any] , lowerCAmelCase__ : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = [] for token in re.findall(self.pat , lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : Any = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(""" """ ) ) return bpe_tokens def UpperCamelCase__ ( self : List[str] , lowerCAmelCase__ : List[str] ): """simple docstring""" return self.encoder.get(lowerCAmelCase__ , self.encoder.get(self.unk_token ) ) def UpperCamelCase__ ( self : int , lowerCAmelCase__ : Optional[int] ): """simple docstring""" return self.decoder.get(lowerCAmelCase__ ) def UpperCamelCase__ ( self : Optional[Any] , lowerCAmelCase__ : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = """""".join(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def UpperCamelCase__ ( self : List[str] , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCAmelCase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __SCREAMING_SNAKE_CASE : int = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __SCREAMING_SNAKE_CASE : Optional[int] = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ ) + """\n""" ) __SCREAMING_SNAKE_CASE : Tuple = 0 with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( F"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." """ Please check that the tokenizer is not corrupted!""" ) __SCREAMING_SNAKE_CASE : List[Any] = token_index writer.write(""" """.join(lowerCAmelCase__ ) + """\n""" ) index += 1 return vocab_file, merge_file def UpperCamelCase__ ( self : str , lowerCAmelCase__ : List[int] , lowerCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __SCREAMING_SNAKE_CASE : List[Any] = [self.cls_token_id] __SCREAMING_SNAKE_CASE : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase__ ( self : Any , lowerCAmelCase__ : List[int] , lowerCAmelCase__ : Optional[List[int]] = None , lowerCAmelCase__ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCamelCase__ ( self : Any , lowerCAmelCase__ : List[int] , lowerCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = [self.sep_token_id] __SCREAMING_SNAKE_CASE : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self : Optional[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : List[Any]=False , **lowerCAmelCase__ : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): __SCREAMING_SNAKE_CASE : int = """ """ + text return (text, kwargs) def UpperCamelCase__ ( self : Optional[Any] , lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] , lowerCAmelCase__ : Optional[int] = None , lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , lowerCAmelCase__ : Optional[int] = None , lowerCAmelCase__ : Optional[bool] = None , ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = super()._pad( encoded_inputs=lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding_strategy=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , ) # Load from model defaults if return_attention_mask is None: __SCREAMING_SNAKE_CASE : Tuple = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: __SCREAMING_SNAKE_CASE : str = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. __SCREAMING_SNAKE_CASE : str = len(encoded_inputs["""global_attention_mask"""] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: __SCREAMING_SNAKE_CASE : Dict = len(lowerCAmelCase__ ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` __SCREAMING_SNAKE_CASE : Union[str, Any] = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": __SCREAMING_SNAKE_CASE : Dict = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
112
1
from math import factorial def UpperCAmelCase_ ( _A , _A ): '''simple docstring''' if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(_A ) // (factorial(_A ) * factorial(n - k )) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', F"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', F"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', F"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
218
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
218
1
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def lowercase ( a__ : List[Any] ) -> List[str]: _UpperCamelCase = 384 _UpperCamelCase = 7 if "tiny" in model_name: _UpperCamelCase = 96 _UpperCamelCase = (2, 2, 6, 2) _UpperCamelCase = (3, 6, 12, 24) elif "small" in model_name: _UpperCamelCase = 96 _UpperCamelCase = (2, 2, 18, 2) _UpperCamelCase = (3, 6, 12, 24) elif "base" in model_name: _UpperCamelCase = 128 _UpperCamelCase = (2, 2, 18, 2) _UpperCamelCase = (4, 8, 16, 32) _UpperCamelCase = 12 _UpperCamelCase = 512 elif "large" in model_name: _UpperCamelCase = 192 _UpperCamelCase = (2, 2, 18, 2) _UpperCamelCase = (6, 12, 24, 48) _UpperCamelCase = 12 _UpperCamelCase = 768 # set label information _UpperCamelCase = 150 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''ade20k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase , repo_type='''dataset''' ) , '''r''' ) ) _UpperCamelCase = {int(__UpperCamelCase ): v for k, v in idalabel.items()} _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = SwinConfig( embed_dim=__UpperCamelCase , depths=__UpperCamelCase , num_heads=__UpperCamelCase , window_size=__UpperCamelCase , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] , ) _UpperCamelCase = UperNetConfig( backbone_config=__UpperCamelCase , auxiliary_in_channels=__UpperCamelCase , num_labels=__UpperCamelCase , idalabel=__UpperCamelCase , labelaid=__UpperCamelCase , ) return config def lowercase ( a__ : Dict ) -> Dict: _UpperCamelCase = [] # fmt: off # stem rename_keys.append(('''backbone.patch_embed.projection.weight''', '''backbone.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.projection.bias''', '''backbone.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''backbone.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''backbone.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.norm1.weight''', F'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.norm1.bias''', F'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table''', F'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index''', F'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight''', F'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias''', F'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.norm2.weight''', F'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.norm2.bias''', F'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight''', F'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias''', F'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight''', F'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((F'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias''', F'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((F'''backbone.stages.{i}.downsample.reduction.weight''', F'''backbone.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((F'''backbone.stages.{i}.downsample.norm.weight''', F'''backbone.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((F'''backbone.stages.{i}.downsample.norm.bias''', F'''backbone.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((F'''backbone.norm{i}.weight''', F'''backbone.hidden_states_norms.stage{i+1}.weight''') ) rename_keys.append((F'''backbone.norm{i}.bias''', F'''backbone.hidden_states_norms.stage{i+1}.bias''') ) # decode head rename_keys.extend( [ ('''decode_head.conv_seg.weight''', '''decode_head.classifier.weight'''), ('''decode_head.conv_seg.bias''', '''decode_head.classifier.bias'''), ('''auxiliary_head.conv_seg.weight''', '''auxiliary_head.classifier.weight'''), ('''auxiliary_head.conv_seg.bias''', '''auxiliary_head.classifier.bias'''), ] ) # fmt: on return rename_keys def lowercase ( a__ : List[Any] , a__ : Union[str, Any] , a__ : int ) -> List[str]: _UpperCamelCase = dct.pop(__UpperCamelCase ) _UpperCamelCase = val def lowercase ( a__ : Union[str, Any] , a__ : List[Any] ) -> Tuple: _UpperCamelCase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): _UpperCamelCase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[:dim, :] _UpperCamelCase = in_proj_bias[: dim] _UpperCamelCase = in_proj_weight[ dim : dim * 2, : ] _UpperCamelCase = in_proj_bias[ dim : dim * 2 ] _UpperCamelCase = in_proj_weight[ -dim :, : ] _UpperCamelCase = in_proj_bias[-dim :] # fmt: on def lowercase ( a__ : List[Any] ) -> Optional[int]: _UpperCamelCase = x.shape _UpperCamelCase = x.reshape(__UpperCamelCase , 4 , in_channel // 4 ) _UpperCamelCase = x[:, [0, 2, 1, 3], :].transpose(1 , 2 ).reshape(__UpperCamelCase , __UpperCamelCase ) return x def lowercase ( a__ : Dict ) -> Optional[int]: _UpperCamelCase = x.shape _UpperCamelCase = x.reshape(__UpperCamelCase , in_channel // 4 , 4 ) _UpperCamelCase = x[:, :, [0, 2, 1, 3]].transpose(1 , 2 ).reshape(__UpperCamelCase , __UpperCamelCase ) return x def lowercase ( a__ : Optional[int] ) -> str: _UpperCamelCase = x.shape[0] _UpperCamelCase = x.reshape(4 , in_channel // 4 ) _UpperCamelCase = x[[0, 2, 1, 3], :].transpose(0 , 1 ).reshape(__UpperCamelCase ) return x def lowercase ( a__ : List[str] ) -> int: _UpperCamelCase = x.shape[0] _UpperCamelCase = x.reshape(in_channel // 4 , 4 ) _UpperCamelCase = x[:, [0, 2, 1, 3]].transpose(0 , 1 ).reshape(__UpperCamelCase ) return x def lowercase ( a__ : Any , a__ : Dict , a__ : int ) -> List[str]: _UpperCamelCase = { '''upernet-swin-tiny''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth''', '''upernet-swin-small''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth''', '''upernet-swin-base''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth''', '''upernet-swin-large''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth''', } _UpperCamelCase = model_name_to_url[model_name] _UpperCamelCase = torch.hub.load_state_dict_from_url(__UpperCamelCase , map_location='''cpu''' , file_name=__UpperCamelCase )[ '''state_dict''' ] for name, param in state_dict.items(): print(__UpperCamelCase , param.shape ) _UpperCamelCase = get_upernet_config(__UpperCamelCase ) _UpperCamelCase = UperNetForSemanticSegmentation(__UpperCamelCase ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _UpperCamelCase = state_dict.pop(__UpperCamelCase ) if "bn" in key: _UpperCamelCase = key.replace('''bn''' , '''batch_norm''' ) _UpperCamelCase = val # rename keys _UpperCamelCase = create_rename_keys(__UpperCamelCase ) for src, dest in rename_keys: rename_key(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) read_in_q_k_v(__UpperCamelCase , config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: _UpperCamelCase = reverse_correct_unfold_reduction_order(__UpperCamelCase ) if "norm" in key: _UpperCamelCase = reverse_correct_unfold_norm_order(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) # verify on image _UpperCamelCase = '''https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg''' _UpperCamelCase = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ).convert('''RGB''' ) _UpperCamelCase = SegformerImageProcessor() _UpperCamelCase = processor(__UpperCamelCase , return_tensors='''pt''' ).pixel_values with torch.no_grad(): _UpperCamelCase = model(__UpperCamelCase ) _UpperCamelCase = outputs.logits print(logits.shape ) print('''First values of logits:''' , logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": _UpperCamelCase = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ) elif model_name == "upernet-swin-small": _UpperCamelCase = torch.tensor( [[-7.1921, -7.1921, -6.9532], [-7.1921, -7.1921, -6.9532], [-7.0908, -7.0908, -6.8534]] ) elif model_name == "upernet-swin-base": _UpperCamelCase = torch.tensor( [[-6.5851, -6.5851, -6.4330], [-6.5851, -6.5851, -6.4330], [-6.4763, -6.4763, -6.3254]] ) elif model_name == "upernet-swin-large": _UpperCamelCase = torch.tensor( [[-7.5297, -7.5297, -7.3802], [-7.5297, -7.5297, -7.3802], [-7.4044, -7.4044, -7.2586]] ) print('''Logits:''' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__UpperCamelCase ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(__UpperCamelCase ) if push_to_hub: print(F'''Pushing model and processor for {model_name} to hub''' ) model.push_to_hub(F'''openmmlab/{model_name}''' ) processor.push_to_hub(F'''openmmlab/{model_name}''' ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""upernet-swin-tiny""", type=str, choices=[F'''upernet-swin-{size}''' for size in ["""tiny""", """small""", """base""", """large"""]], help="""Name of the Swin + UperNet model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) UpperCAmelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
256
"""simple docstring""" from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig a_ = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorch': 'https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json', } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="ernie_m" UpperCamelCase ={"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , UpperCamelCase_ = 25_00_02 , UpperCamelCase_ = 7_68 , UpperCamelCase_ = 12 , UpperCamelCase_ = 12 , UpperCamelCase_ = 30_72 , UpperCamelCase_ = "gelu" , UpperCamelCase_ = 0.1 , UpperCamelCase_ = 0.1 , UpperCamelCase_ = 5_14 , UpperCamelCase_ = 0.0_2 , UpperCamelCase_ = 1 , UpperCamelCase_ = 1E-05 , UpperCamelCase_=None , UpperCamelCase_=False , UpperCamelCase_=0.0 , **UpperCamelCase_ , ) -> List[str]: super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowercase : List[str] = vocab_size __lowercase : Optional[Any] = hidden_size __lowercase : List[Any] = num_hidden_layers __lowercase : Union[str, Any] = num_attention_heads __lowercase : Dict = intermediate_size __lowercase : Any = hidden_act __lowercase : List[str] = hidden_dropout_prob __lowercase : Optional[Any] = attention_probs_dropout_prob __lowercase : Dict = max_position_embeddings __lowercase : Optional[Any] = initializer_range __lowercase : int = layer_norm_eps __lowercase : Optional[int] = classifier_dropout __lowercase : List[Any] = is_decoder __lowercase : List[str] = act_dropout
249
0
'''simple docstring''' import os # Precomputes a list of the 100 first triangular numbers _lowerCamelCase : str = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def __a ( ) ->List[Any]: """simple docstring""" A = os.path.dirname(os.path.realpath(lowercase_ ) ) A = os.path.join(lowercase_ , """words.txt""" ) A = """""" with open(lowercase_ ) as f: A = f.readline() A = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] A = [ word for word in [sum(ord(lowercase_ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowercase_ ) if __name__ == "__main__": print(solution())
362
'''simple docstring''' import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @require_torch def A (self : Any ): A = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) A = load_dataset("""ashraq/esc50""" ) A = dataset["""train"""]["""audio"""][-1]["""array"""] A = audio_classifier(_lowerCAmelCase , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , [{"""score""": 0.501, """label""": """Sound of a dog"""}, {"""score""": 0.499, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def A (self : List[str] ): pass @slow @require_torch def A (self : int ): A = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog A = load_dataset("""ashraq/esc50""" ) A = dataset["""train"""]["""audio"""][-1]["""array"""] A = audio_classifier(_lowerCAmelCase , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ] , ) A = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , [ [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) A = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , [ [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def A (self : Tuple ): pass
337
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase : Optional[Any] = { 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[str] = [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimeSeriesTransformerForPrediction', 'TimeSeriesTransformerModel', 'TimeSeriesTransformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
232
def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : int , _lowerCamelCase : int) -> float: '''simple docstring''' return base * power(_lowerCamelCase , (exponent - 1)) if exponent else 1 if __name__ == "__main__": print('Raise base to the power of exponent using recursion...') lowercase : Optional[int] = int(input('Enter the base: ').strip()) lowercase : Tuple = int(input('Enter the exponent: ').strip()) lowercase : str = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents lowercase : Tuple = 1 / result print(f"{base} to the power of {exponent} is {result}")
232
1
"""simple docstring""" from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase=13 ,__UpperCamelCase=7 ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=99 ,__UpperCamelCase=32 ,__UpperCamelCase=2 ,__UpperCamelCase=4 ,__UpperCamelCase=37 ,__UpperCamelCase="gelu" ,__UpperCamelCase=0.1 ,__UpperCamelCase=0.1 ,__UpperCamelCase=512 ,__UpperCamelCase=16 ,__UpperCamelCase=2 ,__UpperCamelCase=0.02 ,__UpperCamelCase=3 ,__UpperCamelCase=4 ,__UpperCamelCase=None ,) -> int: '''simple docstring''' lowercase_ : int = parent lowercase_ : Union[str, Any] = 13 lowercase_ : Union[str, Any] = 7 lowercase_ : Any = True lowercase_ : Any = True lowercase_ : List[Any] = True lowercase_ : Optional[Any] = True lowercase_ : Tuple = 99 lowercase_ : Tuple = 384 lowercase_ : List[Any] = 2 lowercase_ : Dict = 4 lowercase_ : Tuple = 37 lowercase_ : List[Any] = 'gelu' lowercase_ : List[Any] = 0.1 lowercase_ : int = 0.1 lowercase_ : List[str] = 512 lowercase_ : Optional[int] = 16 lowercase_ : str = 2 lowercase_ : Dict = 0.02 lowercase_ : Optional[Any] = 3 lowercase_ : Optional[Any] = 4 lowercase_ : int = 128 lowercase_ : List[str] = 2 lowercase_ : Any = 9 lowercase_ : List[str] = 1 lowercase_ : str = None def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowercase_ : List[Any] = None if self.use_input_mask: lowercase_ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : Dict = None if self.use_token_type_ids: lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) lowercase_ : Optional[Any] = None lowercase_ : Optional[Any] = None lowercase_ : int = None if self.use_labels: lowercase_ : List[Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) lowercase_ : Optional[Any] = ids_tensor([self.batch_size] ,self.num_choices ) lowercase_ : List[str] = ConvBertConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,return_dict=__UpperCamelCase ,) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ : Optional[int] = TFConvBertModel(config=__UpperCamelCase ) lowercase_ : Union[str, Any] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowercase_ : Tuple = [input_ids, input_mask] lowercase_ : Any = model(__UpperCamelCase ) lowercase_ : Tuple = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Tuple: '''simple docstring''' lowercase_ : Tuple = TFConvBertForMaskedLM(config=__UpperCamelCase ) lowercase_ : Dict = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowercase_ : Any = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ : Any = self.num_labels lowercase_ : List[Any] = TFConvBertForSequenceClassification(config=__UpperCamelCase ) lowercase_ : Any = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowercase_ : List[str] = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Tuple: '''simple docstring''' lowercase_ : str = self.num_choices lowercase_ : Dict = TFConvBertForMultipleChoice(config=__UpperCamelCase ) lowercase_ : Optional[Any] = tf.tile(tf.expand_dims(__UpperCamelCase ,1 ) ,(1, self.num_choices, 1) ) lowercase_ : List[Any] = tf.tile(tf.expand_dims(__UpperCamelCase ,1 ) ,(1, self.num_choices, 1) ) lowercase_ : Dict = tf.tile(tf.expand_dims(__UpperCamelCase ,1 ) ,(1, self.num_choices, 1) ) lowercase_ : List[str] = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } lowercase_ : Optional[int] = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ : List[str] = self.num_labels lowercase_ : Any = TFConvBertForTokenClassification(config=__UpperCamelCase ) lowercase_ : Optional[int] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowercase_ : List[str] = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' lowercase_ : Dict = TFConvBertForQuestionAnswering(config=__UpperCamelCase ) lowercase_ : Tuple = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowercase_ : str = model(__UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Dict = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Optional[Any] = config_and_inputs lowercase_ : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCamelCase ( lowercase_ , lowercase_ , unittest.TestCase ): lowercase = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) lowercase = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) lowercase = False lowercase = False lowercase = False def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' lowercase_ : List[Any] = TFConvBertModelTester(self ) lowercase_ : Any = ConfigTester(self ,config_class=__UpperCamelCase ,hidden_size=37 ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ , lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : List[Any] = True lowercase_ : Optional[Any] = True if hasattr(__UpperCamelCase ,'use_cache' ): lowercase_ : str = True lowercase_ : str = getattr(self.model_tester ,'encoder_seq_length' ,self.model_tester.seq_length ) lowercase_ : Union[str, Any] = getattr(self.model_tester ,'key_length' ,__UpperCamelCase ) for model_class in self.all_model_classes: lowercase_ : Any = self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : str = model_class(__UpperCamelCase ) lowercase_ : Dict = len(model(__UpperCamelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCamelCase ,saved_model=__UpperCamelCase ) lowercase_ : Optional[int] = os.path.join(__UpperCamelCase ,'saved_model' ,'1' ) lowercase_ : Optional[Any] = tf.keras.models.load_model(__UpperCamelCase ) lowercase_ : Dict = model(__UpperCamelCase ) if self.is_encoder_decoder: lowercase_ : Optional[Any] = outputs['encoder_hidden_states'] lowercase_ : List[str] = outputs['encoder_attentions'] else: lowercase_ : int = outputs['hidden_states'] lowercase_ : List[str] = outputs['attentions'] self.assertEqual(len(__UpperCamelCase ) ,__UpperCamelCase ) lowercase_ : Dict = getattr( self.model_tester ,'expected_num_hidden_layers' ,self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__UpperCamelCase ) ,__UpperCamelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) ,[self.model_tester.seq_length, self.model_tester.hidden_size] ,) self.assertEqual(len(__UpperCamelCase ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] ,) @slow def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : str = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) self.assertIsNotNone(__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ , lowercase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Optional[int] = True lowercase_ : str = getattr(self.model_tester ,'decoder_seq_length' ,self.model_tester.seq_length ) lowercase_ : Union[str, Any] = getattr(self.model_tester ,'encoder_seq_length' ,self.model_tester.seq_length ) lowercase_ : str = getattr(self.model_tester ,'key_length' ,__UpperCamelCase ) lowercase_ : Optional[Any] = getattr(self.model_tester ,'key_length' ,__UpperCamelCase ) def check_decoder_attentions_output(__UpperCamelCase ): lowercase_ : str = len(__UpperCamelCase ) self.assertEqual(out_len % 2 ,0 ) lowercase_ : List[Any] = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] ,) def check_encoder_attentions_output(__UpperCamelCase ): lowercase_ : int = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__UpperCamelCase ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] ,) for model_class in self.all_model_classes: lowercase_ : Union[str, Any] = True lowercase_ : Any = False lowercase_ : Optional[int] = model_class(__UpperCamelCase ) lowercase_ : List[str] = model(self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) lowercase_ : Tuple = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states ,__UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: lowercase_ : Optional[Any] = model_class(__UpperCamelCase ) lowercase_ : Any = model(self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) self.assertEqual(config.output_hidden_states ,__UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowercase_ : Tuple = True lowercase_ : Optional[Any] = model_class(__UpperCamelCase ) lowercase_ : Dict = model(self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) self.assertEqual(config.output_hidden_states ,__UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine lowercase_ : Optional[Any] = True lowercase_ : str = True lowercase_ : List[Any] = model_class(__UpperCamelCase ) lowercase_ : str = model(self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) ,len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states ,__UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @require_tf class UpperCamelCase ( unittest.TestCase ): @slow def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : Tuple = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) lowercase_ : Any = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase_ : Dict = model(__UpperCamelCase )[0] lowercase_ : Optional[Any] = [1, 6, 768] self.assertEqual(output.shape ,__UpperCamelCase ) lowercase_ : Tuple = tf.constant( [ [ [-0.0347_5493, -0.468_6034, -0.3063_8832], [0.2263_7248, -0.2698_8646, -0.742_3424], [0.1032_4868, -0.4501_3508, -0.5828_0784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] ,__UpperCamelCase ,atol=1e-4 )
321
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=0.2 ,__UpperCamelCase=0.2 ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Optional[int] = bp_numa lowercase_ : Dict = bp_numa lowercase_ : Tuple = bp_numa lowercase_ : List[Any] = conva_get[:2] lowercase_ : int = conva_get[2] lowercase_ : Dict = size_pa lowercase_ : int = rate_w lowercase_ : Union[str, Any] = rate_t lowercase_ : Dict = [ np.mat(-1 * np.random.rand(self.conva[0] ,self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase_ : Union[str, Any] = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) lowercase_ : Union[str, Any] = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) lowercase_ : str = -2 * np.random.rand(self.conva[1] ) + 1 lowercase_ : Tuple = -2 * np.random.rand(self.num_bpa ) + 1 lowercase_ : Union[str, Any] = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' lowercase_ : int = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(__UpperCamelCase ,'wb' ) as f: pickle.dump(__UpperCamelCase ,__UpperCamelCase ) print(f'''Model saved: {save_path}''' ) @classmethod def _UpperCAmelCase ( cls ,__UpperCamelCase ) -> List[Any]: '''simple docstring''' with open(__UpperCamelCase ,'rb' ) as f: lowercase_ : Any = pickle.load(__UpperCamelCase ) # noqa: S301 lowercase_ : str = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase_ : Union[str, Any] = model_dic.get('size_pooling1' ) lowercase_ : Optional[Any] = model_dic.get('num_bp1' ) lowercase_ : str = model_dic.get('num_bp2' ) lowercase_ : Optional[Any] = model_dic.get('num_bp3' ) lowercase_ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase_ : Optional[int] = model_dic.get('rate_thre' ) # create model instance lowercase_ : Any = CNN(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # modify model parameter lowercase_ : Optional[Any] = model_dic.get('w_conv1' ) lowercase_ : Tuple = model_dic.get('wkj' ) lowercase_ : Union[str, Any] = model_dic.get('vji' ) lowercase_ : Optional[Any] = model_dic.get('thre_conv1' ) lowercase_ : Dict = model_dic.get('thre_bp2' ) lowercase_ : Optional[int] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Any: '''simple docstring''' return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' return round(__UpperCamelCase ,3 ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Tuple: '''simple docstring''' lowercase_ : Dict = convs[0] lowercase_ : Any = convs[1] lowercase_ : Optional[Any] = np.shape(__UpperCamelCase )[0] # get the data slice of original image data, data_focus lowercase_ : Tuple = [] for i_focus in range(0 ,size_data - size_conv + 1 ,__UpperCamelCase ): for j_focus in range(0 ,size_data - size_conv + 1 ,__UpperCamelCase ): lowercase_ : List[Any] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__UpperCamelCase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase_ : Dict = [] lowercase_ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__UpperCamelCase ): lowercase_ : Tuple = [] for i_focus in range(len(__UpperCamelCase ) ): lowercase_ : Optional[int] = ( np.sum(np.multiply(data_focus[i_focus] ,w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__UpperCamelCase ) ) lowercase_ : Optional[int] = np.asmatrix(__UpperCamelCase ).reshape( __UpperCamelCase ,__UpperCamelCase ) data_featuremap.append(__UpperCamelCase ) # expanding the data slice to One dimenssion lowercase_ : Optional[int] = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__UpperCamelCase ) ) lowercase_ : str = np.asarray(__UpperCamelCase ) return focus_list, data_featuremap def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase="average_pool" ) -> Tuple: '''simple docstring''' lowercase_ : Union[str, Any] = len(featuremaps[0] ) lowercase_ : str = int(size_map / size_pooling ) lowercase_ : Optional[int] = [] for i_map in range(len(__UpperCamelCase ) ): lowercase_ : int = featuremaps[i_map] lowercase_ : List[str] = [] for i_focus in range(0 ,__UpperCamelCase ,__UpperCamelCase ): for j_focus in range(0 ,__UpperCamelCase ,__UpperCamelCase ): lowercase_ : List[str] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__UpperCamelCase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__UpperCamelCase ) ) lowercase_ : Dict = np.asmatrix(__UpperCamelCase ).reshape(__UpperCamelCase ,__UpperCamelCase ) featuremap_pooled.append(__UpperCamelCase ) return featuremap_pooled def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Any: '''simple docstring''' lowercase_ : Tuple = [] for i in range(len(__UpperCamelCase ) ): lowercase_ : Optional[Any] = np.shape(data[i] ) lowercase_ : List[str] = data[i].reshape(1 ,shapes[0] * shapes[1] ) lowercase_ : List[str] = data_listed.getA().tolist()[0] data_expanded.extend(__UpperCamelCase ) lowercase_ : int = np.asarray(__UpperCamelCase ) return data_expanded def _UpperCAmelCase ( self ,__UpperCamelCase ) -> int: '''simple docstring''' lowercase_ : Any = np.asarray(__UpperCamelCase ) lowercase_ : Any = np.shape(__UpperCamelCase ) lowercase_ : Optional[Any] = data_mat.reshape(1 ,shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> str: '''simple docstring''' lowercase_ : Any = [] lowercase_ : List[Any] = 0 for i_map in range(__UpperCamelCase ): lowercase_ : List[str] = np.ones((size_map, size_map) ) for i in range(0 ,__UpperCamelCase ,__UpperCamelCase ): for j in range(0 ,__UpperCamelCase ,__UpperCamelCase ): lowercase_ : List[Any] = pd_pool[ i_pool ] lowercase_ : Any = i_pool + 1 lowercase_ : Optional[int] = np.multiply( __UpperCamelCase ,np.multiply(out_map[i_map] ,(1 - out_map[i_map]) ) ) pd_all.append(__UpperCamelCase ) return pd_all def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=bool ) -> Optional[int]: '''simple docstring''' print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(__UpperCamelCase )) ) print((' - - Shape: Teach_Data ', np.shape(__UpperCamelCase )) ) lowercase_ : int = 0 lowercase_ : Tuple = [] lowercase_ : Tuple = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase_ : List[str] = 0 print(f'''-------------Learning Time {rp}--------------''' ) for p in range(len(__UpperCamelCase ) ): # print('------------Learning Image: %d--------------'%p) lowercase_ : int = np.asmatrix(datas_train[p] ) lowercase_ : Any = np.asarray(datas_teach[p] ) lowercase_ , lowercase_ : Tuple = self.convolute( __UpperCamelCase ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) lowercase_ : Any = self.pooling(__UpperCamelCase ,self.size_poolinga ) lowercase_ : Optional[int] = np.shape(__UpperCamelCase ) lowercase_ : Optional[int] = self._expand(__UpperCamelCase ) lowercase_ : int = data_bp_input lowercase_ : Tuple = np.dot(__UpperCamelCase ,self.vji.T ) - self.thre_bpa lowercase_ : Dict = self.sig(__UpperCamelCase ) lowercase_ : int = np.dot(__UpperCamelCase ,self.wkj.T ) - self.thre_bpa lowercase_ : int = self.sig(__UpperCamelCase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase_ : str = np.multiply( (data_teach - bp_outa) ,np.multiply(__UpperCamelCase ,(1 - bp_outa) ) ) lowercase_ : Optional[int] = np.multiply( np.dot(__UpperCamelCase ,self.wkj ) ,np.multiply(__UpperCamelCase ,(1 - bp_outa) ) ) lowercase_ : Any = np.dot(__UpperCamelCase ,self.vji ) lowercase_ : str = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase_ : Dict = pd_conva_pooled.T.getA().tolist() lowercase_ : List[Any] = self._calculate_gradient_from_pool( __UpperCamelCase ,__UpperCamelCase ,shape_featuremapa[0] ,shape_featuremapa[1] ,self.size_poolinga ,) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase_ : Optional[Any] = self._expand_mat(pd_conva_all[k_conv] ) lowercase_ : Dict = self.rate_weight * np.dot(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : List[Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase_ : Dict = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase_ : Optional[int] = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase_ : Any = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase_ : str = self.thre_bpa - pd_k_all * self.rate_thre lowercase_ : Any = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase_ : List[Any] = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase_ : int = rp + 1 lowercase_ : Union[str, Any] = error_count / patterns all_mse.append(__UpperCamelCase ) def draw_error(): lowercase_ : str = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__UpperCamelCase ,'+-' ) plt.plot(__UpperCamelCase ,'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(__UpperCamelCase ,alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' lowercase_ : Union[str, Any] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(__UpperCamelCase )) ) for p in range(len(__UpperCamelCase ) ): lowercase_ : List[Any] = np.asmatrix(datas_test[p] ) lowercase_ , lowercase_ : Optional[Any] = self.convolute( __UpperCamelCase ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) lowercase_ : List[Any] = self.pooling(__UpperCamelCase ,self.size_poolinga ) lowercase_ : List[str] = self._expand(__UpperCamelCase ) lowercase_ : Any = data_bp_input lowercase_ : Optional[Any] = bp_outa * self.vji.T - self.thre_bpa lowercase_ : str = self.sig(__UpperCamelCase ) lowercase_ : List[str] = bp_outa * self.wkj.T - self.thre_bpa lowercase_ : Optional[int] = self.sig(__UpperCamelCase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase_ : List[str] = [list(map(self.do_round ,__UpperCamelCase ) ) for each in produce_out] return np.asarray(__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ : Optional[int] = np.asmatrix(__UpperCamelCase ) lowercase_ , lowercase_ : Union[str, Any] = self.convolute( __UpperCamelCase ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) lowercase_ : Optional[int] = self.pooling(__UpperCamelCase ,self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
321
1
class A_ : def __init__( self : Optional[Any] ,SCREAMING_SNAKE_CASE__ : Tuple ,SCREAMING_SNAKE_CASE__ : Any): __lowerCamelCase : Union[str, Any] = name __lowerCamelCase : Optional[int] = val def __str__( self : str): return F"{self.__class__.__name__}({self.name}, {self.val})" def __lt__( self : str ,SCREAMING_SNAKE_CASE__ : List[str]): return self.val < other.val class A_ : def __init__( self : List[Any] ,SCREAMING_SNAKE_CASE__ : Optional[int]): __lowerCamelCase : Optional[Any] = {} __lowerCamelCase : Union[str, Any] = {} __lowerCamelCase : Any = self.build_heap(SCREAMING_SNAKE_CASE__) def __getitem__( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : str): return self.get_value(SCREAMING_SNAKE_CASE__) def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : Dict): return (idx - 1) // 2 def lowerCAmelCase ( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : List[Any]): return idx * 2 + 1 def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : List[str]): return idx * 2 + 2 def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : Tuple): return self.heap_dict[key] def lowerCAmelCase ( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : Tuple): __lowerCamelCase : Optional[Any] = len(SCREAMING_SNAKE_CASE__) - 1 __lowerCamelCase : Any = self.get_parent_idx(SCREAMING_SNAKE_CASE__) for idx, i in enumerate(SCREAMING_SNAKE_CASE__): __lowerCamelCase : Tuple = idx __lowerCamelCase : Optional[Any] = i.val for i in range(SCREAMING_SNAKE_CASE__ ,-1 ,-1): self.sift_down(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) return array def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : Optional[int] ,SCREAMING_SNAKE_CASE__ : Any): while True: __lowerCamelCase : str = self.get_left_child_idx(SCREAMING_SNAKE_CASE__) # noqa: E741 __lowerCamelCase : List[str] = self.get_right_child_idx(SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = idx if l < len(SCREAMING_SNAKE_CASE__) and array[l] < array[idx]: __lowerCamelCase : Optional[Any] = l if r < len(SCREAMING_SNAKE_CASE__) and array[r] < array[smallest]: __lowerCamelCase : List[str] = r if smallest != idx: __lowerCamelCase , __lowerCamelCase : Union[str, Any] = array[smallest], array[idx] ( ( __lowerCamelCase ) , ( __lowerCamelCase ) , ) : int = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) __lowerCamelCase : Optional[int] = smallest else: break def lowerCAmelCase ( self : str ,SCREAMING_SNAKE_CASE__ : List[Any]): __lowerCamelCase : Optional[Any] = self.get_parent_idx(SCREAMING_SNAKE_CASE__) while p >= 0 and self.heap[p] > self.heap[idx]: __lowerCamelCase , __lowerCamelCase : int = self.heap[idx], self.heap[p] __lowerCamelCase , __lowerCamelCase : Union[str, Any] = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) __lowerCamelCase : str = p __lowerCamelCase : int = self.get_parent_idx(SCREAMING_SNAKE_CASE__) def lowerCAmelCase ( self : List[str]): return self.heap[0] def lowerCAmelCase ( self : int): __lowerCamelCase , __lowerCamelCase : Tuple = self.heap[-1], self.heap[0] __lowerCamelCase , __lowerCamelCase : Tuple = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) __lowerCamelCase : Optional[Any] = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 ,self.heap) return x def lowerCAmelCase ( self : str ,SCREAMING_SNAKE_CASE__ : Union[str, Any]): self.heap.append(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[int] = len(self.heap) - 1 __lowerCamelCase : str = node.val self.sift_up(len(self.heap) - 1) def lowerCAmelCase ( self : Optional[int]): return len(self.heap) == 0 def lowerCAmelCase ( self : int ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : List[Any]): assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" __lowerCamelCase : List[str] = new_value __lowerCamelCase : Dict = new_value self.sift_up(self.idx_of_element[node]) a =Node("""R""", -1) a =Node("""B""", 6) a =Node("""A""", 3) a =Node("""X""", 1) a =Node("""E""", 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array a =MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print("""Min Heap - before decrease key""") for i in my_min_heap.heap: print(i) print("""Min Heap - After decrease key of node [B -> -17]""") my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
73
from __future__ import annotations import time a =list[tuple[int, int]] a =[ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] a =[[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class A_ : def __init__( self : List[str] ,SCREAMING_SNAKE_CASE__ : int ,SCREAMING_SNAKE_CASE__ : int ,SCREAMING_SNAKE_CASE__ : int ,SCREAMING_SNAKE_CASE__ : int ,SCREAMING_SNAKE_CASE__ : Node | None): __lowerCamelCase : Tuple = pos_x __lowerCamelCase : List[str] = pos_y __lowerCamelCase : str = (pos_y, pos_x) __lowerCamelCase : str = goal_x __lowerCamelCase : int = goal_y __lowerCamelCase : List[Any] = parent class A_ : def __init__( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : tuple[int, int] ,SCREAMING_SNAKE_CASE__ : tuple[int, int]): __lowerCamelCase : Any = Node(start[1] ,start[0] ,goal[1] ,goal[0] ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : List[Any] = Node(goal[1] ,goal[0] ,goal[1] ,goal[0] ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Union[str, Any] = [self.start] __lowerCamelCase : List[str] = False def lowerCAmelCase ( self : List[Any]): while self.node_queue: __lowerCamelCase : Any = self.node_queue.pop(0) if current_node.pos == self.target.pos: __lowerCamelCase : Dict = True return self.retrace_path(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Tuple = self.get_successors(SCREAMING_SNAKE_CASE__) for node in successors: self.node_queue.append(SCREAMING_SNAKE_CASE__) if not self.reached: return [self.start.pos] return None def lowerCAmelCase ( self : str ,SCREAMING_SNAKE_CASE__ : Node): __lowerCamelCase : Union[str, Any] = [] for action in delta: __lowerCamelCase : Optional[Any] = parent.pos_x + action[1] __lowerCamelCase : Optional[int] = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(SCREAMING_SNAKE_CASE__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,self.target.pos_y ,self.target.pos_x ,SCREAMING_SNAKE_CASE__)) return successors def lowerCAmelCase ( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : Node | None): __lowerCamelCase : List[Any] = node __lowerCamelCase : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) __lowerCamelCase : int = current_node.parent path.reverse() return path class A_ : def __init__( self : Optional[Any] ,SCREAMING_SNAKE_CASE__ : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : int): __lowerCamelCase : int = BreadthFirstSearch(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = BreadthFirstSearch(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = False def lowerCAmelCase ( self : str): while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: __lowerCamelCase : Any = self.fwd_bfs.node_queue.pop(0) __lowerCamelCase : Any = self.bwd_bfs.node_queue.pop(0) if current_bwd_node.pos == current_fwd_node.pos: __lowerCamelCase : List[str] = True return self.retrace_bidirectional_path( SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = current_bwd_node __lowerCamelCase : int = current_fwd_node __lowerCamelCase : str = { self.fwd_bfs: self.fwd_bfs.get_successors(SCREAMING_SNAKE_CASE__), self.bwd_bfs: self.bwd_bfs.get_successors(SCREAMING_SNAKE_CASE__), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(SCREAMING_SNAKE_CASE__) if not self.reached: return [self.fwd_bfs.start.pos] return None def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : Node ,SCREAMING_SNAKE_CASE__ : Node): __lowerCamelCase : List[Any] = self.fwd_bfs.retrace_path(SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = self.bwd_bfs.retrace_path(SCREAMING_SNAKE_CASE__) bwd_path.pop() bwd_path.reverse() __lowerCamelCase : List[Any] = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() a =(0, 0) a =(len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) a =time.time() a =BreadthFirstSearch(init, goal) a =bfs.search() a =time.time() - start_bfs_time print("""Unidirectional BFS computation time : """, bfs_time) a =time.time() a =BidirectionalBreadthFirstSearch(init, goal) a =bd_bfs.search() a =time.time() - start_bd_bfs_time print("""Bidirectional BFS computation time : """, bd_bfs_time)
73
1
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase = None , ) -> List[Any]: super().__init__() self.register_modules(transformer=__lowercase , vae=__lowercase , scheduler=__lowercase) # create a imagenet -> id dictionary for easier use __UpperCamelCase :Tuple = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split(''','''): __UpperCamelCase :List[Any] = int(__lowercase) __UpperCamelCase :List[str] = dict(sorted(self.labels.items())) def UpperCamelCase__ ( self , __lowercase) -> List[int]: if not isinstance(__lowercase , __lowercase): __UpperCamelCase :Tuple = list(__lowercase) for l in label: if l not in self.labels: raise ValueError( f"""{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.""") return [self.labels[l] for l in label] @torch.no_grad() def __call__( self , __lowercase , __lowercase = 4.0 , __lowercase = None , __lowercase = 50 , __lowercase = "pil" , __lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: __UpperCamelCase :str = len(__lowercase) __UpperCamelCase :List[Any] = self.transformer.config.sample_size __UpperCamelCase :Dict = self.transformer.config.in_channels __UpperCamelCase :List[Any] = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=__lowercase , device=self.device , dtype=self.transformer.dtype , ) __UpperCamelCase :List[Any] = torch.cat([latents] * 2) if guidance_scale > 1 else latents __UpperCamelCase :List[Any] = torch.tensor(__lowercase , device=self.device).reshape(-1) __UpperCamelCase :Optional[int] = torch.tensor([1_000] * batch_size , device=self.device) __UpperCamelCase :Any = torch.cat([class_labels, class_null] , 0) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(__lowercase) for t in self.progress_bar(self.scheduler.timesteps): if guidance_scale > 1: __UpperCamelCase :Tuple = latent_model_input[: len(__lowercase) // 2] __UpperCamelCase :int = torch.cat([half, half] , dim=0) __UpperCamelCase :List[Any] = self.scheduler.scale_model_input(__lowercase , __lowercase) __UpperCamelCase :Any = t if not torch.is_tensor(__lowercase): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) __UpperCamelCase :int = latent_model_input.device.type == '''mps''' if isinstance(__lowercase , __lowercase): __UpperCamelCase :int = torch.floataa if is_mps else torch.floataa else: __UpperCamelCase :Optional[int] = torch.intaa if is_mps else torch.intaa __UpperCamelCase :Tuple = torch.tensor([timesteps] , dtype=__lowercase , device=latent_model_input.device) elif len(timesteps.shape) == 0: __UpperCamelCase :str = timesteps[None].to(latent_model_input.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase :Optional[Any] = timesteps.expand(latent_model_input.shape[0]) # predict noise model_output __UpperCamelCase :Tuple = self.transformer( __lowercase , timestep=__lowercase , class_labels=__lowercase).sample # perform guidance if guidance_scale > 1: __UpperCamelCase :Dict = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] __UpperCamelCase :List[Any] = torch.split(__lowercase , len(__lowercase) // 2 , dim=0) __UpperCamelCase :Union[str, Any] = uncond_eps + guidance_scale * (cond_eps - uncond_eps) __UpperCamelCase :Any = torch.cat([half_eps, half_eps] , dim=0) __UpperCamelCase :List[str] = torch.cat([eps, rest] , dim=1) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: __UpperCamelCase :List[Any] = torch.split(__lowercase , __lowercase , dim=1) else: __UpperCamelCase :str = noise_pred # compute previous image: x_t -> x_t-1 __UpperCamelCase :List[Any] = self.scheduler.step(__lowercase , __lowercase , __lowercase).prev_sample if guidance_scale > 1: __UpperCamelCase :Any = latent_model_input.chunk(2 , dim=0) else: __UpperCamelCase :Optional[int] = latent_model_input __UpperCamelCase :List[str] = 1 / self.vae.config.scaling_factor * latents __UpperCamelCase :Dict = self.vae.decode(__lowercase).sample __UpperCamelCase :Any = (samples / 2 + 0.5).clamp(0 , 1) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 __UpperCamelCase :Any = samples.cpu().permute(0 , 2 , 3 , 1).float().numpy() if output_type == "pil": __UpperCamelCase :Optional[Any] = self.numpy_to_pil(__lowercase) if not return_dict: return (samples,) return ImagePipelineOutput(images=__lowercase)
370
import os import unittest from transformers.models.phobert.tokenization_phobert import VOCAB_FILES_NAMES, PhobertTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : List[Any] = PhobertTokenizer a__ : Union[str, Any] = False def UpperCamelCase__ ( self) -> Optional[int]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase :Union[str, Any] = ['''T@@''', '''i''', '''I''', '''R@@''', '''r''', '''e@@'''] __UpperCamelCase :int = dict(zip(__lowercase , range(len(__lowercase)))) __UpperCamelCase :Dict = ['''#version: 0.2''', '''l à</w>'''] __UpperCamelCase :Any = {'''unk_token''': '''<unk>'''} __UpperCamelCase :int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) __UpperCamelCase :Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: for token in vocab_tokens: fp.write(f"""{token} {vocab_tokens[token]}\n""") with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(__lowercase)) def UpperCamelCase__ ( self , **__lowercase) -> Optional[Any]: kwargs.update(self.special_tokens_map) return PhobertTokenizer.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self , __lowercase) -> int: __UpperCamelCase :List[Any] = '''Tôi là VinAI Research''' __UpperCamelCase :List[str] = '''T<unk> i <unk> <unk> <unk> <unk> <unk> <unk> I Re<unk> e<unk> <unk> <unk> <unk>''' return input_text, output_text def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Dict = PhobertTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) __UpperCamelCase :List[Any] = '''Tôi là VinAI Research''' __UpperCamelCase :List[str] = '''T@@ ô@@ i l@@ à V@@ i@@ n@@ A@@ I R@@ e@@ s@@ e@@ a@@ r@@ c@@ h'''.split() __UpperCamelCase :int = tokenizer.tokenize(__lowercase) print(__lowercase) self.assertListEqual(__lowercase , __lowercase) __UpperCamelCase :Dict = tokens + [tokenizer.unk_token] __UpperCamelCase :Any = [4, 3, 5, 3, 3, 3, 3, 3, 3, 6, 7, 9, 3, 9, 3, 3, 3, 3, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase) , __lowercase)
105
0
from datetime import datetime as dt import os from github import Github lowerCAmelCase__ : Any = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''feature request''', '''new model''', '''wip''', ] def UpperCamelCase__ ( ) -> Optional[int]: snake_case__ : int = Github(os.environ['GITHUB_TOKEN'] ) snake_case__ : List[str] = g.get_repo('huggingface/transformers' ) snake_case__ : Optional[int] = repo.get_issues(state='open' ) for issue in open_issues: snake_case__ : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda A__ : i.created_at , reverse=_a ) snake_case__ : str = comments[0] if len(_a ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='closed' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
143
def lowerCamelCase__ ( _a , _a): return int((input_a, input_a).count(1) != 0) def lowerCamelCase__ ( ): assert or_gate(0 , 0) == 0 assert or_gate(0 , 1) == 1 assert or_gate(1 , 0) == 1 assert or_gate(1 , 1) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
76
0
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class UpperCamelCase__: def __init__( self : Optional[int] , lowerCAmelCase : int , lowerCAmelCase : MutableSequence[float] )-> None: """simple docstring""" if len(lowerCAmelCase ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) UpperCAmelCase = list(lowerCAmelCase ) UpperCAmelCase = degree def __add__( self : Dict , lowerCAmelCase : Polynomial )-> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: UpperCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCAmelCase ) else: UpperCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCAmelCase ) def __sub__( self : List[str] , lowerCAmelCase : Polynomial )-> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Optional[int] )-> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : List[str] , lowerCAmelCase : Polynomial )-> Polynomial: """simple docstring""" UpperCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCAmelCase ) def a__( self : str , lowerCAmelCase : int | float )-> int | float: """simple docstring""" UpperCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : int )-> str: """simple docstring""" UpperCAmelCase = '''''' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] )-> str: """simple docstring""" return self.__str__() def a__( self : Optional[Any] )-> Polynomial: """simple docstring""" UpperCAmelCase = [0] * self.degree for i in range(self.degree ): UpperCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCAmelCase ) def a__( self : Any , lowerCAmelCase : int | float = 0 )-> Polynomial: """simple docstring""" UpperCAmelCase = [0] * (self.degree + 2) UpperCAmelCase = constant for i in range(self.degree + 1 ): UpperCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCAmelCase ) def __eq__( self : str , lowerCAmelCase : object )-> bool: """simple docstring""" if not isinstance(lowerCAmelCase , lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Any , lowerCAmelCase : object )-> bool: """simple docstring""" return not self.__eq__(lowerCAmelCase )
91
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase : Union[str, Any] = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """PoolFormerOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : List[Any] = ["""PoolFormerFeatureExtractor"""] _lowercase : Any = ["""PoolFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Tuple = [ """POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PoolFormerForImageClassification""", """PoolFormerModel""", """PoolFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _lowercase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
91
1
"""simple docstring""" import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class _UpperCAmelCase ( __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =FlaxAutoencoderKL @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = 4 __snake_case : List[Any] = 3 __snake_case : str = (32, 32) __snake_case : Union[str, Any] = jax.random.PRNGKey(0 ) __snake_case : Union[str, Any] = jax.random.uniform(a_ , ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } __snake_case : Optional[int] = self.dummy_input return init_dict, inputs_dict
102
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __A ( SCREAMING_SNAKE_CASE_ ): def __init__( self , a__ , a__=13 , a__=7 , a__=True , a__=True , a__=True , a__=True , a__=99 , a__=32 , a__=5 , a__=4 , a__=37 , a__="gelu" , a__=0.1 , a__=0.1 , a__=512 , a__=16 , a__=2 , a__=0.0_2 , a__=False , a__=True , a__="None" , a__=3 , a__=4 , a__=None , ): _lowerCAmelCase : Dict = parent _lowerCAmelCase : str = batch_size _lowerCAmelCase : List[Any] = seq_length _lowerCAmelCase : Dict = is_training _lowerCAmelCase : Dict = use_input_mask _lowerCAmelCase : int = use_token_type_ids _lowerCAmelCase : int = use_labels _lowerCAmelCase : Optional[int] = vocab_size _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : Tuple = num_hidden_layers _lowerCAmelCase : Dict = num_attention_heads _lowerCAmelCase : Union[str, Any] = intermediate_size _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : List[str] = attention_probs_dropout_prob _lowerCAmelCase : List[str] = max_position_embeddings _lowerCAmelCase : List[str] = type_vocab_size _lowerCAmelCase : Tuple = type_sequence_label_size _lowerCAmelCase : List[Any] = initializer_range _lowerCAmelCase : Union[str, Any] = num_labels _lowerCAmelCase : Optional[Any] = num_choices _lowerCAmelCase : Tuple = relative_attention _lowerCAmelCase : Tuple = position_biased_input _lowerCAmelCase : Dict = pos_att_type _lowerCAmelCase : Any = scope def __A ( self ): _lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : Optional[Any] = None if self.use_input_mask: _lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _lowerCAmelCase : str = None if self.use_token_type_ids: _lowerCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : Union[str, Any] = None _lowerCAmelCase : Union[str, Any] = None _lowerCAmelCase : Any = None if self.use_labels: _lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) _lowerCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __A ( self ): return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def __A ( self , a__ ): self.parent.assertListEqual(list(result.loss.size() ) , [] ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Union[str, Any] = DebertaVaModel(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : List[Any] = model(a__ , attention_mask=a__ , token_type_ids=a__ )[0] _lowerCAmelCase : List[Any] = model(a__ , token_type_ids=a__ )[0] _lowerCAmelCase : Any = model(a__ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : List[str] = DebertaVaForMaskedLM(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Optional[Any] = model(a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : int = self.num_labels _lowerCAmelCase : int = DebertaVaForSequenceClassification(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Optional[Any] = model(a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(a__ ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : str = DebertaVaForTokenClassification(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Optional[Any] = model(a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Any = DebertaVaForQuestionAnswering(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Dict = model( a__ , attention_mask=a__ , token_type_ids=a__ , start_positions=a__ , end_positions=a__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Union[str, Any] = DebertaVaForMultipleChoice(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Optional[int] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCAmelCase : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCAmelCase : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCAmelCase : List[str] = model( a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __A ( self ): _lowerCAmelCase : Tuple = self.prepare_config_and_inputs() ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCAmelCase : Union[str, Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __A ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Union[str, Any] = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) _UpperCamelCase : str = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase : Optional[Any] = True _UpperCamelCase : List[Any] = False _UpperCamelCase : List[Any] = False _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = False def __A ( self ): _lowerCAmelCase : Optional[Any] = DebertaVaModelTester(self ) _lowerCAmelCase : Any = ConfigTester(self , config_class=a__ , hidden_size=37 ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*a__ ) def __A ( self ): _lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*a__ ) def __A ( self ): _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*a__ ) def __A ( self ): _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*a__ ) def __A ( self ): _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*a__ ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*a__ ) @slow def __A ( self ): for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Tuple = DebertaVaModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) @require_torch @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def __A ( self ): pass @slow def __A ( self ): _lowerCAmelCase : Union[str, Any] = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _lowerCAmelCase : Dict = torch.tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) _lowerCAmelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(a__ , attention_mask=a__ )[0] # compare the actual values for a slice. _lowerCAmelCase : str = torch.tensor( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , a__ , atol=1e-4 ) , F"{output[:, 1:4, 1:4]}" )
44
0
from __future__ import annotations UpperCamelCase__ = 1.6021E-19 # units = C def _a ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , ): if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif conductivity < 0: raise ValueError("Conductivity cannot be negative" ) elif electron_conc < 0: raise ValueError("Electron concentration cannot be negative" ) elif mobility < 0: raise ValueError("mobility cannot be negative" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
368
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class a__ : def __init__( self , _A , _A=1_3 , _A=3_0 , _A=2 , _A=3 , _A=True , _A=True , _A=3_2 , _A=2 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=1_0 , _A=0.02 , _A=3 , _A=None , _A=2 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = scope __lowerCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) __lowerCAmelCase = (image_size // patch_size) ** 2 __lowerCAmelCase = num_patches + 2 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = self.get_config() return config, pixel_values, labels def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_A , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = TFDeiTModel(config=_A ) __lowerCAmelCase = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = TFDeiTForMaskedImageModeling(config=_A ) __lowerCAmelCase = model(_A ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowerCAmelCase = 1 __lowerCAmelCase = TFDeiTForMaskedImageModeling(_A ) __lowerCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase = model(_A ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = self.type_sequence_label_size __lowerCAmelCase = TFDeiTForImageClassification(_A ) __lowerCAmelCase = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCAmelCase = 1 __lowerCAmelCase = TFDeiTForImageClassification(_A ) __lowerCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class a__ ( snake_case__ , snake_case__ , unittest.TestCase ): _a : Optional[Any] = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) _a : Optional[Any] = ( { """feature-extraction""": TFDeiTModel, """image-classification""": (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) _a : str = False _a : str = False _a : List[str] = False _a : Optional[int] = False def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFDeiTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=3_7 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) __lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , tf.keras.layers.Dense ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(_A ) __lowerCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A=False ): """simple docstring""" __lowerCAmelCase = super()._prepare_for_class(_A , _A , return_labels=_A ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFDeiTModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def _a ( ): __lowerCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class a__ ( unittest.TestCase ): @cached_property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ) __lowerCAmelCase = self.default_image_processor __lowerCAmelCase = prepare_img() __lowerCAmelCase = image_processor(images=_A , return_tensors="tf" ) # forward pass __lowerCAmelCase = model(**_A ) # verify the logits __lowerCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _A ) __lowerCAmelCase = tf.constant([-1.02_66, 0.19_12, -1.28_61] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) )
102
0
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( '''The `inpainting.py` script is outdated. Please use directly `from diffusers import''' ''' StableDiffusionInpaintPipeline` instead.''' )
119
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ (__A ): def __init__( self : str , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : List[Any]=13 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : str=True , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[str]=99 , lowerCAmelCase_ : int=32 , lowerCAmelCase_ : List[str]=5 , lowerCAmelCase_ : Optional[int]=4 , lowerCAmelCase_ : str=37 , lowerCAmelCase_ : List[Any]="gelu" , lowerCAmelCase_ : Dict=0.1 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : List[Any]=512 , lowerCAmelCase_ : Optional[int]=16 , lowerCAmelCase_ : Union[str, Any]=2 , lowerCAmelCase_ : List[str]=0.0_2 , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]="None" , lowerCAmelCase_ : List[Any]=3 , lowerCAmelCase_ : Optional[int]=4 , lowerCAmelCase_ : int=None , ) -> Dict: UpperCAmelCase_ : Dict = parent UpperCAmelCase_ : Union[str, Any] = batch_size UpperCAmelCase_ : Optional[Any] = seq_length UpperCAmelCase_ : List[Any] = is_training UpperCAmelCase_ : Optional[int] = use_input_mask UpperCAmelCase_ : int = use_token_type_ids UpperCAmelCase_ : Any = use_labels UpperCAmelCase_ : Optional[int] = vocab_size UpperCAmelCase_ : Any = hidden_size UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : List[Any] = intermediate_size UpperCAmelCase_ : int = hidden_act UpperCAmelCase_ : Optional[Any] = hidden_dropout_prob UpperCAmelCase_ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase_ : Any = max_position_embeddings UpperCAmelCase_ : Union[str, Any] = type_vocab_size UpperCAmelCase_ : Union[str, Any] = type_sequence_label_size UpperCAmelCase_ : Tuple = initializer_range UpperCAmelCase_ : int = num_labels UpperCAmelCase_ : Optional[Any] = num_choices UpperCAmelCase_ : List[str] = relative_attention UpperCAmelCase_ : List[Any] = position_biased_input UpperCAmelCase_ : Dict = pos_att_type UpperCAmelCase_ : Optional[Any] = scope def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Dict: UpperCAmelCase_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ : Tuple = None if self.use_input_mask: UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) UpperCAmelCase_ : Optional[Any] = None if self.use_token_type_ids: UpperCAmelCase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : List[str] = None UpperCAmelCase_ : Union[str, Any] = None if self.use_labels: UpperCAmelCase_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase_ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase_ : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : List[str] = self.get_config() UpperCAmelCase_ : int = 300 return config def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase_ : int ) -> List[Any]: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Any , lowerCAmelCase_ : Any , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[Any] ) -> List[Any]: UpperCAmelCase_ : Optional[Any] = DebertaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] UpperCAmelCase_ : Optional[int] = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] UpperCAmelCase_ : Optional[Any] = model(lowerCAmelCase_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Optional[int] ) -> List[Any]: UpperCAmelCase_ : Union[str, Any] = DebertaForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[Any] ) -> Optional[Any]: UpperCAmelCase_ : Any = self.num_labels UpperCAmelCase_ : List[Any] = DebertaForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple ) -> str: UpperCAmelCase_ : Optional[int] = self.num_labels UpperCAmelCase_ : Optional[int] = DebertaForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : int , lowerCAmelCase_ : str ) -> List[Any]: UpperCAmelCase_ : Dict = DebertaForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Any = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : Tuple = config_and_inputs UpperCAmelCase_ : List[Any] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ (__A , __A , unittest.TestCase ): __magic_name__ = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) __magic_name__ = ( { '''feature-extraction''': DebertaModel, '''fill-mask''': DebertaForMaskedLM, '''question-answering''': DebertaForQuestionAnswering, '''text-classification''': DebertaForSequenceClassification, '''token-classification''': DebertaForTokenClassification, '''zero-shot''': DebertaForSequenceClassification, } if is_torch_available() else {} ) __magic_name__ = True __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def _SCREAMING_SNAKE_CASE ( self : Dict ) -> int: UpperCAmelCase_ : int = DebertaModelTester(self ) UpperCAmelCase_ : Any = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[Any]: self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Optional[Any]: UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[int]: UpperCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : int ) -> Union[str, Any]: UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[int] = DebertaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase_ (unittest.TestCase ): @unittest.skip(reason="Model not available yet" ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: pass @slow def _SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: UpperCAmelCase_ : Optional[int] = DebertaModel.from_pretrained("microsoft/deberta-base" ) UpperCAmelCase_ : List[Any] = torch.tensor([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) UpperCAmelCase_ : Tuple = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] # compare the actual values for a slice. UpperCAmelCase_ : Tuple = torch.tensor( [[[-0.5_9_8_6, -0.8_0_5_5, -0.8_4_6_2], [1.4_4_8_4, -0.9_3_4_8, -0.8_0_5_9], [0.3_1_2_3, 0.0_0_3_2, -1.4_1_3_1]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1e-4 ) , f"""{output[:, 1:4, 1:4]}""" )
268
0
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets _a = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' _a = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' _a = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _A ( UpperCamelCase_ : Optional[int], UpperCamelCase_ : int) -> str: '''simple docstring''' return float((preds == labels).mean()) def _A ( UpperCamelCase_ : Dict, UpperCamelCase_ : Any) -> List[str]: '''simple docstring''' __lowercase = simple_accuracy(UpperCamelCase_, UpperCamelCase_) __lowercase = float(fa_score(y_true=UpperCamelCase_, y_pred=UpperCamelCase_)) return { "accuracy": acc, "f1": fa, } def _A ( UpperCamelCase_ : Optional[Any], UpperCamelCase_ : int) -> Optional[int]: '''simple docstring''' __lowercase = np.array(UpperCamelCase_) __lowercase = np.array(UpperCamelCase_) __lowercase = en_sentvecs.shape[0] # mean centering __lowercase = en_sentvecs - np.mean(UpperCamelCase_, axis=0) __lowercase = in_sentvecs - np.mean(UpperCamelCase_, axis=0) __lowercase = cdist(UpperCamelCase_, UpperCamelCase_, "cosine") __lowercase = np.array(range(UpperCamelCase_)) __lowercase = sim.argsort(axis=1)[:, :10] __lowercase = np.any(preds == actual[:, None], axis=1) return float(matches.mean()) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def _lowercase ( self : List[str] ): if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]" ) return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "predictions": datasets.Value("int64" ) if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32" ) ), "references": datasets.Value("int64" ) if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32" ) ), } ), codebase_urls=[], reference_urls=[], format="numpy" if self.config_name != "cvit-mkb-clsr" else None, ) def _lowercase ( self : Dict, UpperCAmelCase__ : str, UpperCAmelCase__ : Tuple ): if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(UpperCAmelCase__, UpperCAmelCase__ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(UpperCAmelCase__, UpperCAmelCase__ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(UpperCAmelCase__, UpperCAmelCase__ )} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]" )
144
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : Optional[Any] ): __lowercase = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) __lowercase = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]], dtype=tf.intaa, ) # J'aime le camembert !" __lowercase = model(UpperCAmelCase__ )["last_hidden_state"] __lowercase = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape, UpperCAmelCase__ ) # compare the actual values for a slice. __lowercase = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]], dtype=tf.floataa, ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy(), expected_slice.numpy(), atol=1E-4 ) )
144
1
'''simple docstring''' def UpperCamelCase_ ( _UpperCAmelCase : bytes ) -> str: """simple docstring""" return "".join([hex(_UpperCAmelCase )[2:].zfill(2 ).upper() for byte in list(_UpperCAmelCase )] ) def UpperCamelCase_ ( _UpperCAmelCase : str ) -> bytes: """simple docstring""" if (len(_UpperCAmelCase ) % 2) != 0: raise ValueError( "Base16 encoded data is invalid:\nData does not have an even number of hex digits." ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(_UpperCAmelCase ) <= set("0123456789ABCDEF" ): raise ValueError( "Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters." ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(_UpperCAmelCase ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
31
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class a__( unittest.TestCase ): @slow def lowercase_ ( self : List[Any] ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: a : Optional[int] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Dict = TFAutoModel.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Tuple = AutoModel.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : str ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: a : List[str] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Union[str, Any] = TFAutoModelForPreTraining.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : List[Any] = AutoModelForPreTraining.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : int ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Union[str, Any] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : List[Any] = TFAutoModelForCausalLM.from_pretrained(__snake_case , from_pt=__snake_case ) a , a : Any = TFAutoModelForCausalLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Dict = AutoModelForCausalLM.from_pretrained(__snake_case , from_tf=__snake_case ) a , a : Tuple = AutoModelForCausalLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : Any ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Tuple = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : List[str] = TFAutoModelWithLMHead.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Dict = AutoModelWithLMHead.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : Optional[int] ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : List[str] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Optional[Any] = TFAutoModelForMaskedLM.from_pretrained(__snake_case , from_pt=__snake_case ) a , a : Optional[int] = TFAutoModelForMaskedLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : str = AutoModelForMaskedLM.from_pretrained(__snake_case , from_tf=__snake_case ) a , a : Tuple = AutoModelForMaskedLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : int ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Optional[Any] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : str = TFAutoModelForSeqaSeqLM.from_pretrained(__snake_case , from_pt=__snake_case ) a , a : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Dict = AutoModelForSeqaSeqLM.from_pretrained(__snake_case , from_tf=__snake_case ) a , a : str = AutoModelForSeqaSeqLM.from_pretrained( __snake_case , output_loading_info=__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : Optional[Any] ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: a : Tuple = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : List[Any] = TFAutoModelForSequenceClassification.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Dict = AutoModelForSequenceClassification.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) @slow def lowercase_ ( self : str ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: a : Optional[Any] = AutoConfig.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : int = TFAutoModelForQuestionAnswering.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) a : Tuple = AutoModelForQuestionAnswering.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsNotNone(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowercase_ ( self : Tuple ): a : List[Any] = TFAutoModelWithLMHead.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=__snake_case ) , 1_44_10 ) a : Optional[int] = AutoModelWithLMHead.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=__snake_case ) , 1_44_10 ) def lowercase_ ( self : Any ): a : int = TFAutoModelWithLMHead.from_pretrained(__snake_case , from_pt=__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=__snake_case ) , 1_44_10 ) a : Optional[Any] = AutoModelWithLMHead.from_pretrained(__snake_case , from_tf=__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=__snake_case ) , 1_44_10 )
297
0
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( snake_case : str )-> list: '''simple docstring''' return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(snake_case ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__("""doctest""").testmod()
298
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ : def __init__( self : Optional[Any] , snake_case__ : Optional[Any] , snake_case__ : Dict=1_3 , snake_case__ : List[str]=7 , snake_case__ : Union[str, Any]=True , snake_case__ : Tuple=True , snake_case__ : Optional[int]=True , snake_case__ : Any=True , snake_case__ : Any=9_9 , snake_case__ : List[Any]=1_6 , snake_case__ : Any=3_6 , snake_case__ : Union[str, Any]=6 , snake_case__ : Tuple=6 , snake_case__ : List[str]=6 , snake_case__ : List[str]=3_7 , snake_case__ : Dict="gelu" , snake_case__ : int=0.1 , snake_case__ : List[Any]=0.1 , snake_case__ : List[str]=5_1_2 , snake_case__ : Dict=1_6 , snake_case__ : str=2 , snake_case__ : Optional[Any]=0.02 , snake_case__ : List[str]=3 , snake_case__ : Any=4 , snake_case__ : int=None , ): '''simple docstring''' UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : int = seq_length UpperCAmelCase__ : List[str] = is_training UpperCAmelCase__ : Union[str, Any] = use_input_mask UpperCAmelCase__ : Optional[Any] = use_token_type_ids UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Any = embedding_size UpperCAmelCase__ : List[str] = hidden_size UpperCAmelCase__ : List[Any] = num_hidden_layers UpperCAmelCase__ : int = num_hidden_groups UpperCAmelCase__ : Union[str, Any] = num_attention_heads UpperCAmelCase__ : List[str] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : List[Any] = hidden_dropout_prob UpperCAmelCase__ : Tuple = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Any = type_vocab_size UpperCAmelCase__ : Union[str, Any] = type_sequence_label_size UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Tuple = num_labels UpperCAmelCase__ : List[str] = num_choices UpperCAmelCase__ : Union[str, Any] = scope def __a ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Optional[int] = None if self.use_input_mask: UpperCAmelCase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Optional[int] = None if self.use_token_type_ids: UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : Any = None if self.use_labels: UpperCAmelCase__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __a ( self : Any ): '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def __a ( self : List[Any] , snake_case__ : Optional[Any] , snake_case__ : List[Any] , snake_case__ : Optional[Any] , snake_case__ : Tuple , snake_case__ : List[str] , snake_case__ : List[Any] , snake_case__ : List[str] ): '''simple docstring''' UpperCAmelCase__ : str = AlbertModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Union[str, Any] = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase__ : Optional[Any] = model(snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase__ : Optional[int] = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __a ( self : Dict , snake_case__ : List[Any] , snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : int ): '''simple docstring''' UpperCAmelCase__ : List[str] = AlbertForPreTraining(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Dict = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , sentence_order_label=snake_case__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def __a ( self : Union[str, Any] , snake_case__ : Any , snake_case__ : List[str] , snake_case__ : List[str] , snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : Optional[Any] , snake_case__ : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = AlbertForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Union[str, Any] = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self : Optional[int] , snake_case__ : Optional[int] , snake_case__ : Optional[Any] , snake_case__ : List[str] , snake_case__ : Any , snake_case__ : Optional[Any] , snake_case__ : List[str] , snake_case__ : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = AlbertForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : List[str] = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __a ( self : Dict , snake_case__ : Optional[Any] , snake_case__ : List[str] , snake_case__ : Optional[int] , snake_case__ : Any , snake_case__ : Any , snake_case__ : List[str] , snake_case__ : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.num_labels UpperCAmelCase__ : int = AlbertForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : int = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self : str , snake_case__ : Optional[Any] , snake_case__ : str , snake_case__ : Optional[Any] , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] , snake_case__ : Dict , snake_case__ : Tuple ): '''simple docstring''' UpperCAmelCase__ : str = self.num_labels UpperCAmelCase__ : Any = AlbertForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : List[str] = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __a ( self : Any , snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : Dict , snake_case__ : int , snake_case__ : List[str] , snake_case__ : Union[str, Any] , snake_case__ : Dict ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.num_choices UpperCAmelCase__ : Optional[Any] = AlbertForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ : Any = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ : Optional[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ : Tuple = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __a ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Optional[Any] = config_and_inputs UpperCAmelCase__ : Optional[int] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): SCREAMING_SNAKE_CASE_ =( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ =( { '''feature-extraction''': AlbertModel, '''fill-mask''': AlbertForMaskedLM, '''question-answering''': AlbertForQuestionAnswering, '''text-classification''': AlbertForSequenceClassification, '''token-classification''': AlbertForTokenClassification, '''zero-shot''': AlbertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ =True def __a ( self : Tuple , snake_case__ : int , snake_case__ : Tuple , snake_case__ : Optional[int]=False ): '''simple docstring''' UpperCAmelCase__ : List[str] = super()._prepare_for_class(snake_case__ , snake_case__ , return_labels=snake_case__ ) if return_labels: if model_class in get_values(snake_case__ ): UpperCAmelCase__ : List[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=snake_case__ ) UpperCAmelCase__ : int = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case__ ) return inputs_dict def __a ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = AlbertModelTester(self ) UpperCAmelCase__ : Any = ConfigTester(self , config_class=snake_case__ , hidden_size=3_7 ) def __a ( self : Dict ): '''simple docstring''' self.config_tester.run_common_tests() def __a ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case__ ) def __a ( self : str ): '''simple docstring''' UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case__ ) def __a ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case__ ) def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case__ ) def __a ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case__ ) def __a ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ : Dict = type self.model_tester.create_and_check_model(*snake_case__ ) @slow def __a ( self : Union[str, Any] ): '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : List[str] = AlbertModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class lowerCAmelCase__ ( unittest.TestCase ): @slow def __a ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[str] = AlbertModel.from_pretrained("albert-base-v2" ) UpperCAmelCase__ : Dict = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase__ : List[str] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCAmelCase__ : int = model(snake_case__ , attention_mask=snake_case__ )[0] UpperCAmelCase__ : Dict = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , snake_case__ , atol=1e-4 ) )
298
1
"""simple docstring""" class snake_case : def __init__( self : List[str] , UpperCamelCase__ : Optional[int])-> str: '''simple docstring''' __lowerCAmelCase: Tuple = arr.split(",") def lowercase_ ( self : Union[str, Any])-> int: '''simple docstring''' __lowerCAmelCase: Any = [int(self.array[0])] * len(self.array) __lowerCAmelCase: Any = [int(self.array[0])] * len(self.array) for i in range(1 , len(self.array)): __lowerCAmelCase: List[str] = max( int(self.array[i]) + sum_value[i - 1] , int(self.array[i])) __lowerCAmelCase: Union[str, Any] = max(sum_value[i] , rear[i - 1]) return rear[len(self.array) - 1] if __name__ == "__main__": __A = input("please input some numbers:") __A = SubArray(whole_array) __A = array.solve_sub_array() print(("the results is:", re))
217
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() __snake_case =logging.get_logger(__name__) def a_ ( lowerCamelCase : Any ): lowerCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCAmelCase = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCAmelCase = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCAmelCase = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCAmelCase = key.replace(f'''patch_embed{idx}''' , f'''patch_embeddings.{int(lowerCamelCase )-1}''' ) if "norm" in key: lowerCAmelCase = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCAmelCase = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCAmelCase = key.replace(f'''layer_norm{idx}''' , f'''layer_norm.{int(lowerCamelCase )-1}''' ) if "layer_norm1" in key: lowerCAmelCase = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCAmelCase = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCAmelCase = key[key.find('block' ) + len('block' )] lowerCAmelCase = key.replace(f'''block{idx}''' , f'''block.{int(lowerCamelCase )-1}''' ) if "attn.q" in key: lowerCAmelCase = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCAmelCase = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCAmelCase = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCAmelCase = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCAmelCase = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCAmelCase = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCAmelCase = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCAmelCase = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCAmelCase = key[key.find('linear_c' ) + len('linear_c' )] lowerCAmelCase = key.replace(f'''linear_c{idx}''' , f'''linear_c.{int(lowerCamelCase )-1}''' ) if "bot_conv" in key: lowerCAmelCase = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCAmelCase = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCAmelCase = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCAmelCase = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCAmelCase = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCAmelCase = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCAmelCase = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCAmelCase = key.replace('module.last_layer_depth' , 'head.head' ) lowerCAmelCase = value return new_state_dict def a_ ( lowerCamelCase : List[str] , lowerCamelCase : str ): # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCAmelCase = state_dict.pop(f'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) lowerCAmelCase = state_dict.pop(f'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict lowerCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] lowerCAmelCase = kv_bias[: config.hidden_sizes[i]] lowerCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] lowerCAmelCase = kv_bias[config.hidden_sizes[i] :] def a_ ( ): lowerCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCAmelCase = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ) return image @torch.no_grad() def a_ ( lowerCamelCase : List[Any] , lowerCamelCase : Tuple , lowerCamelCase : Optional[Any]=False , lowerCamelCase : List[str]=None ): lowerCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCAmelCase = GLPNImageProcessor() # prepare image lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=lowerCamelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCAmelCase = torch.load(lowerCamelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCAmelCase = rename_keys(lowerCamelCase ) # key and value matrices need special treatment read_in_k_v(lowerCamelCase , lowerCamelCase ) # create HuggingFace model and load state dict lowerCAmelCase = GLPNForDepthEstimation(lowerCamelCase ) model.load_state_dict(lowerCamelCase ) model.eval() # forward pass lowerCAmelCase = model(lowerCamelCase ) lowerCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCAmelCase = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCAmelCase = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(f'''Unknown model name: {model_name}''' ) lowerCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , lowerCamelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(lowerCamelCase , lowerCamelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=lowerCamelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(lowerCamelCase , lowerCamelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=lowerCamelCase , ) if __name__ == "__main__": __snake_case =argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) __snake_case =parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
4
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Tuple = { """salesforce/blip2-opt-2.7b""": """https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json""", } class a__ ( __A ): """simple docstring""" __UpperCamelCase : Optional[Any] = 'blip_2_vision_model' def __init__(self , __lowercase=14_08 , __lowercase=61_44 , __lowercase=39 , __lowercase=16 , __lowercase=2_24 , __lowercase=14 , __lowercase="gelu" , __lowercase=0.0_0_0_0_1 , __lowercase=0.0 , __lowercase=1e-10 , __lowercase=True , **__lowercase , ): super().__init__(**__lowercase ) __lowerCAmelCase = hidden_size __lowerCAmelCase = intermediate_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = patch_size __lowerCAmelCase = image_size __lowerCAmelCase = initializer_range __lowerCAmelCase = attention_dropout __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = hidden_act __lowerCAmelCase = qkv_bias @classmethod def _snake_case (cls , __lowercase , **__lowercase ): cls._set_token_in_kwargs(__lowercase ) __lowerCAmelCase , __lowerCAmelCase = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from Blip2Config if config_dict.get('''model_type''' ) == "blip-2": __lowerCAmelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a__ ( __A ): """simple docstring""" __UpperCamelCase : Dict = 'blip_2_qformer' def __init__(self , __lowercase=3_05_22 , __lowercase=7_68 , __lowercase=12 , __lowercase=12 , __lowercase=30_72 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_12 , __lowercase=0.0_2 , __lowercase=1e-12 , __lowercase=0 , __lowercase="absolute" , __lowercase=2 , __lowercase=14_08 , **__lowercase , ): super().__init__(pad_token_id=__lowercase , **__lowercase ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = hidden_act __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = position_embedding_type __lowerCAmelCase = cross_attention_frequency __lowerCAmelCase = encoder_hidden_size @classmethod def _snake_case (cls , __lowercase , **__lowercase ): cls._set_token_in_kwargs(__lowercase ) __lowerCAmelCase , __lowerCAmelCase = cls.get_config_dict(__lowercase , **__lowercase ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get('''model_type''' ) == "blip-2": __lowerCAmelCase = config_dict['''qformer_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a__ ( __A ): """simple docstring""" __UpperCamelCase : Dict = 'blip-2' __UpperCamelCase : Optional[Any] = True def __init__(self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=32 , **__lowercase ): super().__init__(**__lowercase ) if vision_config is None: __lowerCAmelCase = {} logger.info('''vision_config is None. initializing the Blip2VisionConfig with default values.''' ) if qformer_config is None: __lowerCAmelCase = {} logger.info('''qformer_config is None. Initializing the Blip2QFormerConfig with default values.''' ) if text_config is None: __lowerCAmelCase = {} logger.info('''text_config is None. Initializing the text config with default values (`OPTConfig`).''' ) __lowerCAmelCase = BlipaVisionConfig(**__lowercase ) __lowerCAmelCase = BlipaQFormerConfig(**__lowercase ) __lowerCAmelCase = text_config['''model_type'''] if '''model_type''' in text_config else '''opt''' __lowerCAmelCase = CONFIG_MAPPING[text_model_type](**__lowercase ) __lowerCAmelCase = self.text_config.tie_word_embeddings __lowerCAmelCase = self.text_config.is_encoder_decoder __lowerCAmelCase = num_query_tokens __lowerCAmelCase = self.vision_config.hidden_size __lowerCAmelCase = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __lowerCAmelCase = 1.0 __lowerCAmelCase = 0.0_2 @classmethod def _snake_case (cls , __lowercase , __lowercase , __lowercase , **__lowercase , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__lowercase , ) def _snake_case (self ): __lowerCAmelCase = copy.deepcopy(self.__dict__ ) __lowerCAmelCase = self.vision_config.to_dict() __lowerCAmelCase = self.qformer_config.to_dict() __lowerCAmelCase = self.text_config.to_dict() __lowerCAmelCase = self.__class__.model_type return output
9
'''simple docstring''' from ..utils import DummyObject, requires_backends class a__ ( metaclass=__A ): """simple docstring""" __UpperCamelCase : int = ['torch', 'scipy'] def __init__(self , *__lowercase , **__lowercase ): requires_backends(self , ['''torch''', '''scipy'''] ) @classmethod def _snake_case (cls , *__lowercase , **__lowercase ): requires_backends(cls , ['''torch''', '''scipy'''] ) @classmethod def _snake_case (cls , *__lowercase , **__lowercase ): requires_backends(cls , ['''torch''', '''scipy'''] )
9
1
"""simple docstring""" def lowercase_ ( _lowerCamelCase: int , _lowerCamelCase: int ) -> List[str]: '''simple docstring''' if b == 0: return 1 if (b % 2) == 0: return actual_power(_lowerCamelCase , int(b / 2 ) ) * actual_power(_lowerCamelCase , int(b / 2 ) ) else: return a * actual_power(_lowerCamelCase , int(b / 2 ) ) * actual_power(_lowerCamelCase , int(b / 2 ) ) def lowercase_ ( _lowerCamelCase: int , _lowerCamelCase: int ) -> float: '''simple docstring''' if b < 0: return 1 / actual_power(_lowerCamelCase , _lowerCamelCase ) return actual_power(_lowerCamelCase , _lowerCamelCase ) if __name__ == "__main__": print(power(-2, -3))
135
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer __A = logging.get_logger(__name__) __A = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __A = { '''vocab_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt''' ), '''google/realm-orqa-nq-openqa''': '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-nq-reader''': '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt''', '''google/realm-orqa-wq-openqa''': '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-wq-reader''': '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-openqa''': ( '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-reader''': ( '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-openqa''': ( '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-reader''': ( '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json''' ), }, } __A = { '''google/realm-cc-news-pretrained-embedder''': 512, '''google/realm-cc-news-pretrained-encoder''': 512, '''google/realm-cc-news-pretrained-scorer''': 512, '''google/realm-cc-news-pretrained-openqa''': 512, '''google/realm-orqa-nq-openqa''': 512, '''google/realm-orqa-nq-reader''': 512, '''google/realm-orqa-wq-openqa''': 512, '''google/realm-orqa-wq-reader''': 512, } __A = { '''google/realm-cc-news-pretrained-embedder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-encoder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-scorer''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-reader''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-reader''': {'''do_lower_case''': True}, } class _snake_case ( a__ ): snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_INIT_CONFIGURATION snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Any=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : int=True , UpperCAmelCase : List[Any]="[UNK]" , UpperCAmelCase : Tuple="[SEP]" , UpperCAmelCase : List[str]="[PAD]" , UpperCAmelCase : Tuple="[CLS]" , UpperCAmelCase : List[Any]="[MASK]" , UpperCAmelCase : str=True , UpperCAmelCase : Union[str, Any]=None , **UpperCAmelCase : Any , ): super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) __lowerCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , UpperCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , UpperCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , UpperCAmelCase ) != tokenize_chinese_chars ): __lowerCamelCase : str = getattr(UpperCAmelCase , normalizer_state.pop("type" ) ) __lowerCamelCase : Any = do_lower_case __lowerCamelCase : List[Any] = strip_accents __lowerCamelCase : Optional[Any] = tokenize_chinese_chars __lowerCamelCase : int = normalizer_class(**UpperCAmelCase ) __lowerCamelCase : List[Any] = do_lower_case def lowerCamelCase__ ( self : Union[str, Any] , UpperCAmelCase : Dict , **UpperCAmelCase : int ): __lowerCamelCase : Optional[int] = PaddingStrategy.MAX_LENGTH __lowerCamelCase : List[Any] = text __lowerCamelCase : Optional[int] = kwargs.pop("text_pair" , UpperCAmelCase ) __lowerCamelCase : List[Any] = kwargs.pop("return_tensors" , UpperCAmelCase ) __lowerCamelCase : Dict = { "input_ids": [], "attention_mask": [], "token_type_ids": [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: __lowerCamelCase : List[str] = batch_text_pair[idx] else: __lowerCamelCase : Optional[int] = None __lowerCamelCase : List[str] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) __lowerCamelCase : Union[str, Any] = encoded_candidates.get("input_ids" ) __lowerCamelCase : Optional[int] = encoded_candidates.get("attention_mask" ) __lowerCamelCase : int = encoded_candidates.get("token_type_ids" ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) __lowerCamelCase : Union[str, Any] = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def lowerCamelCase__ ( self : Optional[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ): __lowerCamelCase : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase__ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ): __lowerCamelCase : Tuple = [self.sep_token_id] __lowerCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ): __lowerCamelCase : Any = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
135
1
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : torch.FloatTensor class a (_lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" @register_to_config def __init__( self : Optional[int] , lowerCamelCase : int = 65536 , lowerCamelCase : Optional[int] = None , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , lowerCamelCase : int = 0 , lowerCamelCase : str = "fourier" , lowerCamelCase : bool = True , lowerCamelCase : bool = False , lowerCamelCase : float = 0.0 , lowerCamelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , lowerCamelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , lowerCamelCase : Tuple[str] = "UNetMidBlock1D" , lowerCamelCase : str = None , lowerCamelCase : Tuple[int] = (32, 32, 64) , lowerCamelCase : str = None , lowerCamelCase : int = 8 , lowerCamelCase : int = 1 , lowerCamelCase : bool = False , ) -> str: super().__init__() __snake_case : int = sample_size # time if time_embedding_type == "fourier": __snake_case : str = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=lowerCamelCase , log=lowerCamelCase , flip_sin_to_cos=lowerCamelCase ) __snake_case : List[str] = 2 * block_out_channels[0] elif time_embedding_type == "positional": __snake_case : str = Timesteps( block_out_channels[0] , flip_sin_to_cos=lowerCamelCase , downscale_freq_shift=lowerCamelCase ) __snake_case : Union[str, Any] = block_out_channels[0] if use_timestep_embedding: __snake_case : Optional[int] = block_out_channels[0] * 4 __snake_case : Union[str, Any] = TimestepEmbedding( in_channels=lowerCamelCase , time_embed_dim=lowerCamelCase , act_fn=lowerCamelCase , out_dim=block_out_channels[0] , ) __snake_case : List[Any] = nn.ModuleList([] ) __snake_case : Tuple = None __snake_case : Optional[Any] = nn.ModuleList([] ) __snake_case : Union[str, Any] = None # down __snake_case : List[Any] = in_channels for i, down_block_type in enumerate(lowerCamelCase ): __snake_case : str = output_channel __snake_case : Dict = block_out_channels[i] if i == 0: input_channel += extra_in_channels __snake_case : Any = i == len(lowerCamelCase ) - 1 __snake_case : List[str] = get_down_block( lowerCamelCase , num_layers=lowerCamelCase , in_channels=lowerCamelCase , out_channels=lowerCamelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(lowerCamelCase ) # mid __snake_case : str = get_mid_block( lowerCamelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=lowerCamelCase , add_downsample=lowerCamelCase , ) # up __snake_case : List[str] = list(reversed(lowerCamelCase ) ) __snake_case : str = reversed_block_out_channels[0] if out_block_type is None: __snake_case : Tuple = out_channels else: __snake_case : List[Any] = block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase ): __snake_case : Any = output_channel __snake_case : Dict = ( reversed_block_out_channels[i + 1] if i < len(lowerCamelCase ) - 1 else final_upsample_channels ) __snake_case : List[str] = i == len(lowerCamelCase ) - 1 __snake_case : str = get_up_block( lowerCamelCase , num_layers=lowerCamelCase , in_channels=lowerCamelCase , out_channels=lowerCamelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(lowerCamelCase ) __snake_case : List[str] = output_channel # out __snake_case : Any = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) __snake_case : Any = get_out_block( out_block_type=lowerCamelCase , num_groups_out=lowerCamelCase , embed_dim=block_out_channels[0] , out_channels=lowerCamelCase , act_fn=lowerCamelCase , fc_dim=block_out_channels[-1] // 4 , ) def __snake_case ( self : List[Any] , lowerCamelCase : torch.FloatTensor , lowerCamelCase : Union[torch.Tensor, float, int] , lowerCamelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: __snake_case : List[Any] = timestep if not torch.is_tensor(lowerCamelCase ): __snake_case : Dict = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(lowerCamelCase ) and len(timesteps.shape ) == 0: __snake_case : List[str] = timesteps[None].to(sample.device ) __snake_case : List[Any] = self.time_proj(lowerCamelCase ) if self.config.use_timestep_embedding: __snake_case : str = self.time_mlp(lowerCamelCase ) else: __snake_case : Any = timestep_embed[..., None] __snake_case : Optional[Any] = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) __snake_case : List[Any] = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down __snake_case : List[str] = () for downsample_block in self.down_blocks: __snake_case : str = downsample_block(hidden_states=lowerCamelCase , temb=lowerCamelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: __snake_case : List[str] = self.mid_block(lowerCamelCase , lowerCamelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): __snake_case : Optional[int] = down_block_res_samples[-1:] __snake_case : Optional[int] = down_block_res_samples[:-1] __snake_case : Any = upsample_block(lowerCamelCase , res_hidden_states_tuple=lowerCamelCase , temb=lowerCamelCase ) # 5. post-process if self.out_block: __snake_case : int = self.out_block(lowerCamelCase , lowerCamelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=lowerCamelCase )
369
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == r: for j in range(__lowerCamelCase ): print(data[j] , end=" " ) print(" " ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location __snake_case : Union[str, Any] = arr[i] combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index + 1 , __lowerCamelCase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): # A temporary array to store all combination one by one __snake_case : Union[str, Any] = [0] * r # Print all combination using temporary array 'data[]' combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , 0 , __lowerCamelCase , 0 ) if __name__ == "__main__": # Driver code to check the function above _snake_case : List[str] = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
134
0
def __UpperCamelCase ( _A : list ) ->Dict: """simple docstring""" if len(_UpperCAmelCase ) < 2: return collection def circle_sort_util(_A : list , _A : int , _A : int ) -> bool: lowerCamelCase_ =False if low == high: return swapped lowerCamelCase_ =low lowerCamelCase_ =high while left < right: if collection[left] > collection[right]: lowerCamelCase_ =( collection[right], collection[left], ) lowerCamelCase_ =True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: lowerCamelCase_ =( collection[right + 1], collection[left], ) lowerCamelCase_ =True lowerCamelCase_ =low + int((high - low) / 2 ) lowerCamelCase_ =circle_sort_util(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase_ =circle_sort_util(_UpperCAmelCase , mid + 1 , _UpperCAmelCase ) return swapped or left_swap or right_swap lowerCamelCase_ =True while is_not_sorted is True: lowerCamelCase_ =circle_sort_util(_UpperCAmelCase , 0 , len(_UpperCAmelCase ) - 1 ) return collection if __name__ == "__main__": __A : List[Any] = input('Enter numbers separated by a comma:\n').strip() __A : Any = [int(item) for item in user_input.split(',')] print(circle_sort(unsorted))
154
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() __A =logging.get_logger(__name__) def a ( _UpperCAmelCase : int ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = SwinConfig( embed_dim=1_92 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['''stage2''', '''stage3''', '''stage4'''] , ) __UpperCAmelCase : Optional[int] = DetaConfig( backbone_config=_UpperCAmelCase , num_queries=9_00 , encoder_ffn_dim=20_48 , decoder_ffn_dim=20_48 , num_feature_levels=5 , assign_first_stage=_UpperCAmelCase , with_box_refine=_UpperCAmelCase , two_stage=_UpperCAmelCase , ) # set labels __UpperCAmelCase : Optional[int] = '''huggingface/label-files''' if "o365" in model_name: __UpperCAmelCase : Tuple = 3_66 __UpperCAmelCase : List[str] = '''object365-id2label.json''' else: __UpperCAmelCase : Any = 91 __UpperCAmelCase : int = '''coco-detection-id2label.json''' __UpperCAmelCase : Optional[int] = num_labels __UpperCAmelCase : List[str] = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) __UpperCAmelCase : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} __UpperCAmelCase : Optional[int] = idalabel __UpperCAmelCase : Tuple = {v: k for k, v in idalabel.items()} return config def a ( _UpperCAmelCase : Dict ): '''simple docstring''' __UpperCAmelCase : List[str] = [] # stem # fmt: off rename_keys.append(('''backbone.0.body.patch_embed.proj.weight''', '''model.backbone.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.proj.bias''', '''model.backbone.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.weight''', '''model.backbone.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.bias''', '''model.backbone.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias') ) if i < 3: rename_keys.append((f'backbone.0.body.layers.{i}.downsample.reduction.weight', f'model.backbone.model.encoder.layers.{i}.downsample.reduction.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.weight', f'model.backbone.model.encoder.layers.{i}.downsample.norm.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.bias', f'model.backbone.model.encoder.layers.{i}.downsample.norm.bias') ) rename_keys.append(('''backbone.0.body.norm1.weight''', '''model.backbone.model.hidden_states_norms.stage2.weight''') ) rename_keys.append(('''backbone.0.body.norm1.bias''', '''model.backbone.model.hidden_states_norms.stage2.bias''') ) rename_keys.append(('''backbone.0.body.norm2.weight''', '''model.backbone.model.hidden_states_norms.stage3.weight''') ) rename_keys.append(('''backbone.0.body.norm2.bias''', '''model.backbone.model.hidden_states_norms.stage3.bias''') ) rename_keys.append(('''backbone.0.body.norm3.weight''', '''model.backbone.model.hidden_states_norms.stage4.weight''') ) rename_keys.append(('''backbone.0.body.norm3.bias''', '''model.backbone.model.hidden_states_norms.stage4.bias''') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight', f'model.encoder.layers.{i}.self_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias', f'model.encoder.layers.{i}.self_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.weight', f'model.encoder.layers.{i}.self_attn.attention_weights.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.bias', f'model.encoder.layers.{i}.self_attn.attention_weights.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.weight', f'model.encoder.layers.{i}.self_attn.value_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.bias', f'model.encoder.layers.{i}.self_attn.value_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.weight', f'model.encoder.layers.{i}.self_attn.output_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.bias', f'model.encoder.layers.{i}.self_attn.output_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.weight', f'model.encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'model.encoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'model.encoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'model.encoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'model.encoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'model.encoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'model.encoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'model.encoder.layers.{i}.final_layer_norm.bias') ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.weight', f'model.decoder.layers.{i}.encoder_attn.attention_weights.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.bias', f'model.decoder.layers.{i}.encoder_attn.attention_weights.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.weight', f'model.decoder.layers.{i}.encoder_attn.value_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.bias', f'model.decoder.layers.{i}.encoder_attn.value_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.weight', f'model.decoder.layers.{i}.encoder_attn.output_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.bias', f'model.decoder.layers.{i}.encoder_attn.output_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.weight', f'model.decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'model.decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'model.decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'model.decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.weight', f'model.decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.bias', f'model.decoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'model.decoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'model.decoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'model.decoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'model.decoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'model.decoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'model.decoder.layers.{i}.final_layer_norm.bias') ) # fmt: on return rename_keys def a ( _UpperCAmelCase : Tuple , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : Optional[int] = dct.pop(_UpperCAmelCase ) __UpperCAmelCase : List[Any] = val def a ( _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any] ): '''simple docstring''' __UpperCAmelCase : List[str] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __UpperCAmelCase : str = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __UpperCAmelCase : List[str] = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight' ) __UpperCAmelCase : List[Any] = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __UpperCAmelCase : Dict = in_proj_weight[:dim, :] __UpperCAmelCase : List[str] = in_proj_bias[: dim] __UpperCAmelCase : str = in_proj_weight[ dim : dim * 2, : ] __UpperCAmelCase : Any = in_proj_bias[ dim : dim * 2 ] __UpperCAmelCase : Tuple = in_proj_weight[ -dim :, : ] __UpperCAmelCase : int = in_proj_bias[-dim :] # fmt: on def a ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Any ): '''simple docstring''' __UpperCAmelCase : int = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention __UpperCAmelCase : List[str] = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) __UpperCAmelCase : Tuple = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict __UpperCAmelCase : Union[str, Any] = in_proj_weight[:hidden_size, :] __UpperCAmelCase : List[Any] = in_proj_bias[:hidden_size] __UpperCAmelCase : int = in_proj_weight[ hidden_size : hidden_size * 2, : ] __UpperCAmelCase : str = in_proj_bias[hidden_size : hidden_size * 2] __UpperCAmelCase : Tuple = in_proj_weight[-hidden_size:, :] __UpperCAmelCase : Optional[Any] = in_proj_bias[-hidden_size:] def a ( ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __UpperCAmelCase : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def a ( _UpperCAmelCase : str , _UpperCAmelCase : List[Any] , _UpperCAmelCase : List[str] ): '''simple docstring''' __UpperCAmelCase : Tuple = get_deta_config(_UpperCAmelCase ) # load original state dict if model_name == "deta-swin-large": __UpperCAmelCase : Dict = hf_hub_download(repo_id='''nielsr/deta-checkpoints''' , filename='''adet_swin_ft.pth''' ) elif model_name == "deta-swin-large-o365": __UpperCAmelCase : Any = hf_hub_download(repo_id='''jozhang97/deta-swin-l-o365''' , filename='''deta_swin_pt_o365.pth''' ) else: raise ValueError(f'Model name {model_name} not supported' ) __UpperCAmelCase : str = torch.load(_UpperCAmelCase , map_location='''cpu''' )['''model'''] # original state dict for name, param in state_dict.items(): print(_UpperCAmelCase , param.shape ) # rename keys __UpperCAmelCase : int = create_rename_keys(_UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_swin_q_k_v(_UpperCAmelCase , config.backbone_config ) read_in_decoder_q_k_v(_UpperCAmelCase , _UpperCAmelCase ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: __UpperCAmelCase : Optional[Any] = state_dict.pop(_UpperCAmelCase ) __UpperCAmelCase : Union[str, Any] = val if "input_proj" in key: __UpperCAmelCase : Union[str, Any] = state_dict.pop(_UpperCAmelCase ) __UpperCAmelCase : List[str] = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: __UpperCAmelCase : Union[str, Any] = state_dict.pop(_UpperCAmelCase ) __UpperCAmelCase : Union[str, Any] = val # finally, create HuggingFace model and load state dict __UpperCAmelCase : Union[str, Any] = DetaForObjectDetection(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() __UpperCAmelCase : Optional[int] = '''cuda''' if torch.cuda.is_available() else '''cpu''' model.to(_UpperCAmelCase ) # load image processor __UpperCAmelCase : str = DetaImageProcessor(format='''coco_detection''' ) # verify our conversion on image __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : Optional[int] = processor(images=_UpperCAmelCase , return_tensors='''pt''' ) __UpperCAmelCase : List[Any] = encoding['''pixel_values'''] __UpperCAmelCase : List[str] = model(pixel_values.to(_UpperCAmelCase ) ) # verify logits print('''Logits:''' , outputs.logits[0, :3, :3] ) print('''Boxes:''' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": __UpperCAmelCase : str = torch.tensor( [[-7.63_08, -2.84_85, -5.37_37], [-7.20_37, -4.55_05, -4.80_27], [-7.29_43, -4.26_11, -4.66_17]] ) __UpperCAmelCase : Union[str, Any] = torch.tensor([[0.49_87, 0.49_69, 0.99_99], [0.25_49, 0.54_98, 0.48_05], [0.54_98, 0.27_57, 0.05_69]] ) elif model_name == "deta-swin-large-o365": __UpperCAmelCase : Optional[Any] = torch.tensor( [[-8.01_22, -3.57_20, -4.97_17], [-8.15_47, -3.68_86, -4.63_89], [-7.66_10, -3.61_94, -5.01_34]] ) __UpperCAmelCase : str = torch.tensor([[0.25_23, 0.55_49, 0.48_81], [0.77_15, 0.41_49, 0.46_01], [0.55_03, 0.27_53, 0.05_75]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(_UpperCAmelCase ) , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(_UpperCAmelCase ) , atol=1e-4 ) print('''Everything ok!''' ) if pytorch_dump_folder_path: # Save model and processor logger.info(f'Saving PyTorch model and processor to {pytorch_dump_folder_path}...' ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) # Push to hub if push_to_hub: print('''Pushing model and processor to hub...''' ) model.push_to_hub(f'jozhang97/{model_name}' ) processor.push_to_hub(f'jozhang97/{model_name}' ) if __name__ == "__main__": __A =argparse.ArgumentParser() parser.add_argument( "--model_name", type=str, default="deta-swin-large", choices=["deta-swin-large", "deta-swin-large-o365"], help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model.", ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __A =parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
226
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore _snake_case = "\nHuman: <<task>>\n\nAssistant: " _snake_case = "huggingface-tools/default-prompts" _snake_case = {"chat": "chat_prompt_template.txt", "run": "run_prompt_template.txt"} def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase="run" ): '''simple docstring''' if prompt_or_repo_id is None: _lowerCAmelCase : str = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , _lowerCamelCase ) is not None: return prompt_or_repo_id _lowerCAmelCase : Dict = cached_file( _lowerCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(_lowerCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
359
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def A ( _lowerCamelCase = "laptop" ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = F"https://www.amazon.in/laptop/s?k={product}" _lowerCAmelCase : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } _lowerCAmelCase : Optional[int] = BeautifulSoup(requests.get(_lowerCamelCase , headers=_lowerCamelCase ).text ) # Initialize a Pandas dataframe with the column titles _lowerCAmelCase : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: _lowerCAmelCase : Any = item.ha.text _lowerCAmelCase : List[str] = "https://www.amazon.in/" + item.ha.a["href"] _lowerCAmelCase : Any = item.find("span" , attrs={"class": "a-offscreen"} ).text try: _lowerCAmelCase : List[str] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: _lowerCAmelCase : str = "Not available" try: _lowerCAmelCase : Optional[Any] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: _lowerCAmelCase : Optional[Any] = "" try: _lowerCAmelCase : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: _lowerCAmelCase : Optional[Any] = float("nan" ) except AttributeError: pass _lowerCAmelCase : Any = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCAmelCase : List[str] = " " _lowerCAmelCase : Tuple = " " data_frame.index += 1 return data_frame if __name__ == "__main__": _snake_case = "headphones" get_amazon_product_data(product).to_csv(f'''Amazon Product Data for {product}.csv''')
300
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowercase : Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class __snake_case ( lowerCAmelCase ): _a : Optional[Any]= ["pixel_values"] def __init__( self ,snake_case = True ,snake_case = None ,snake_case = PILImageResampling.BICUBIC ,snake_case = True ,snake_case = None ,snake_case = True ,snake_case = 1 / 255 ,snake_case = True ,snake_case = None ,snake_case = None ,snake_case = True ,**snake_case ,): '''simple docstring''' super().__init__(**snake_case ) lowercase : List[str] = size if size is not None else {"""shortest_edge""": 224} lowercase : Optional[int] = get_size_dict(snake_case ,default_to_square=snake_case ) lowercase : List[Any] = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} lowercase : Optional[Any] = get_size_dict(snake_case ,default_to_square=snake_case ,param_name="""crop_size""" ) lowercase : Dict = do_resize lowercase : Dict = size lowercase : str = resample lowercase : Tuple = do_center_crop lowercase : Tuple = crop_size lowercase : Union[str, Any] = do_rescale lowercase : List[Any] = rescale_factor lowercase : Any = do_normalize lowercase : Any = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowercase : Dict = image_std if image_std is not None else OPENAI_CLIP_STD lowercase : Dict = do_convert_rgb def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case = PILImageResampling.BICUBIC ,snake_case = None ,**snake_case ,): '''simple docstring''' lowercase : Union[str, Any] = get_size_dict(snake_case ,default_to_square=snake_case ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) lowercase : Optional[Any] = get_resize_output_image_size(snake_case ,size=size["""shortest_edge"""] ,default_to_square=snake_case ) return resize(snake_case ,size=snake_case ,resample=snake_case ,data_format=snake_case ,**snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case = None ,**snake_case ,): '''simple docstring''' lowercase : Dict = get_size_dict(snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(snake_case ,size=(size["""height"""], size["""width"""]) ,data_format=snake_case ,**snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case = None ,**snake_case ,): '''simple docstring''' return rescale(snake_case ,scale=snake_case ,data_format=snake_case ,**snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case ,snake_case = None ,**snake_case ,): '''simple docstring''' return normalize(snake_case ,mean=snake_case ,std=snake_case ,data_format=snake_case ,**snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = None ,snake_case = ChannelDimension.FIRST ,**snake_case ,): '''simple docstring''' lowercase : Dict = do_resize if do_resize is not None else self.do_resize lowercase : Optional[int] = size if size is not None else self.size lowercase : Dict = get_size_dict(snake_case ,param_name="""size""" ,default_to_square=snake_case ) lowercase : Any = resample if resample is not None else self.resample lowercase : Any = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase : Tuple = crop_size if crop_size is not None else self.crop_size lowercase : str = get_size_dict(snake_case ,param_name="""crop_size""" ,default_to_square=snake_case ) lowercase : str = do_rescale if do_rescale is not None else self.do_rescale lowercase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase : str = do_normalize if do_normalize is not None else self.do_normalize lowercase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowercase : List[Any] = image_std if image_std is not None else self.image_std lowercase : List[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowercase : List[Any] = make_list_of_images(snake_case ) if not valid_images(snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowercase : Optional[int] = [convert_to_rgb(snake_case ) for image in images] # All transformations expect numpy arrays. lowercase : Optional[Any] = [to_numpy_array(snake_case ) for image in images] if do_resize: lowercase : str = [self.resize(image=snake_case ,size=snake_case ,resample=snake_case ) for image in images] if do_center_crop: lowercase : Tuple = [self.center_crop(image=snake_case ,size=snake_case ) for image in images] if do_rescale: lowercase : Union[str, Any] = [self.rescale(image=snake_case ,scale=snake_case ) for image in images] if do_normalize: lowercase : Optional[Any] = [self.normalize(image=snake_case ,mean=snake_case ,std=snake_case ) for image in images] lowercase : List[Any] = [to_channel_dimension_format(snake_case ,snake_case ) for image in images] lowercase : int = {"""pixel_values""": images} return BatchFeature(data=snake_case ,tensor_type=snake_case )
20
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : int = {"vocab_file": "vocab.txt"} _UpperCAmelCase : str = { "vocab_file": { "openbmb/cpm-ant-10b": "https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt", }, } _UpperCAmelCase : Optional[Any] = { "openbmb/cpm-ant-10b": 1_024, } def A ( lowercase ) -> Dict: '''simple docstring''' UpperCamelCase = collections.OrderedDict() with open(lowercase , 'r' , encoding='utf-8' ) as reader: UpperCamelCase = reader.readlines() for index, token in enumerate(lowercase ): UpperCamelCase = token.rstrip('\n' ) UpperCamelCase = index return vocab class lowercase ( _SCREAMING_SNAKE_CASE ): def __init__( self , A_ , A_="<unk>" , A_=200 ) -> Dict: """simple docstring""" UpperCamelCase = vocab UpperCamelCase = unk_token UpperCamelCase = max_input_chars_per_word def __UpperCamelCase ( self , A_ ) -> List[Any]: """simple docstring""" UpperCamelCase = list(A_ ) if len(A_ ) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase = 0 UpperCamelCase = [] while start < len(A_ ): UpperCamelCase = len(A_ ) UpperCamelCase = None while start < end: UpperCamelCase = ''.join(chars[start:end] ) if substr in self.vocab: UpperCamelCase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(A_ ) UpperCamelCase = end return sub_tokens class lowercase ( _SCREAMING_SNAKE_CASE ): __lowercase : List[str] = VOCAB_FILES_NAMES __lowercase : Dict = PRETRAINED_VOCAB_FILES_MAP __lowercase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Any = ["input_ids", "attention_mask"] __lowercase : Tuple = False def __init__( self , A_ , A_="<d>" , A_="</d>" , A_="<s>" , A_="</s>" , A_="<pad>" , A_="<unk>" , A_="</n>" , A_="</_>" , A_="left" , **A_ , ) -> Tuple: """simple docstring""" requires_backends(self , ['jieba'] ) super().__init__( bod_token=A_ , eod_token=A_ , bos_token=A_ , eos_token=A_ , pad_token=A_ , unk_token=A_ , line_token=A_ , space_token=A_ , padding_side=A_ , **A_ , ) UpperCamelCase = bod_token UpperCamelCase = eod_token UpperCamelCase = load_vocab(A_ ) UpperCamelCase = self.encoder[space_token] UpperCamelCase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda A_ : x[1] ) ) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def __UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" return self.encoder[self.bod_token] @property def __UpperCamelCase ( self ) -> Tuple: """simple docstring""" return self.encoder[self.eod_token] @property def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" return self.encoder["\n"] @property def __UpperCamelCase ( self ) -> int: """simple docstring""" return len(self.encoder ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def __UpperCamelCase ( self , A_ ) -> List[Any]: """simple docstring""" UpperCamelCase = [] for x in jieba.cut(A_ , cut_all=A_ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(A_ ) ) return output_tokens def __UpperCamelCase ( self , A_ , **A_ ) -> Dict: """simple docstring""" UpperCamelCase = [i for i in token_ids if i >= 0] UpperCamelCase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(A_ , **A_ ) def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" return token in self.encoder def __UpperCamelCase ( self , A_ ) -> str: """simple docstring""" return "".join(A_ ) def __UpperCamelCase ( self , A_ ) -> Tuple: """simple docstring""" return self.encoder.get(A_ , self.encoder.get(self.unk_token ) ) def __UpperCamelCase ( self , A_ ) -> Optional[Any]: """simple docstring""" return self.decoder.get(A_ , self.unk_token ) def __UpperCamelCase ( self , A_ , A_ = None ) -> Tuple[str]: """simple docstring""" if os.path.isdir(A_ ): UpperCamelCase = os.path.join( A_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: UpperCamelCase = (filename_prefix + '-' if filename_prefix else '') + save_directory UpperCamelCase = 0 if " " in self.encoder: UpperCamelCase = self.encoder[' '] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase = self.encoder['\n'] del self.encoder["\n"] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda A_ : x[1] ) ) with open(A_ , 'w' , encoding='utf-8' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ' Please check that the vocabulary is not corrupted!' ) UpperCamelCase = token_index writer.write(token + '\n' ) index += 1 return (vocab_file,) def __UpperCamelCase ( self , A_ , A_ = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def __UpperCamelCase ( self , A_ , A_ = None , A_ = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is not None: return [1] + ([0] * len(A_ )) + [1] + ([0] * len(A_ )) return [1] + ([0] * len(A_ ))
222
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : List[str] =logging.get_logger(__name__) __lowerCAmelCase : int ={ "google/switch-base-8": "https://huggingface.co/google/switch-base-8/blob/main/config.json", } class UpperCAmelCase ( UpperCamelCase__ ): __lowercase = """switch_transformers""" __lowercase = ["""past_key_values"""] __lowercase = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self :str , lowercase_ :int=3_21_28 , lowercase_ :List[str]=7_68 , lowercase_ :str=64 , lowercase_ :int=20_48 , lowercase_ :Union[str, Any]=64 , lowercase_ :Any=12 , lowercase_ :int=3 , lowercase_ :int=12 , lowercase_ :int=3 , lowercase_ :int=12 , lowercase_ :Optional[Any]=8 , lowercase_ :Dict=False , lowercase_ :Optional[int]=0.0_1 , lowercase_ :Any="float32" , lowercase_ :Optional[int]=False , lowercase_ :Any=32 , lowercase_ :Tuple=1_28 , lowercase_ :Optional[int]=0.1 , lowercase_ :Tuple=1E-6 , lowercase_ :List[Any]=0.0_0_1 , lowercase_ :Any=0.0_0_1 , lowercase_ :int=1.0 , lowercase_ :int="relu" , lowercase_ :List[str]=True , lowercase_ :List[str]=False , lowercase_ :int=True , lowercase_ :Any=0 , lowercase_ :Optional[Any]=1 , **lowercase_ :Union[str, Any] , )-> Optional[int]: A__ = vocab_size A__ = d_model A__ = d_kv A__ = d_ff A__ = num_sparse_encoder_layers A__ = num_layers A__ = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry A__ = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: A__ = self.num_layers // self.num_sparse_encoder_layers else: A__ = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: A__ = self.num_decoder_layers // self.num_sparse_decoder_layers else: A__ = self.num_decoder_layers # HACK: this will create 0 sparse layers A__ = num_heads A__ = num_experts A__ = expert_capacity A__ = router_bias A__ = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) A__ = router_dtype A__ = router_ignore_padding_tokens A__ = relative_attention_num_buckets A__ = relative_attention_max_distance A__ = dropout_rate A__ = layer_norm_epsilon A__ = initializer_factor A__ = feed_forward_proj A__ = use_cache A__ = add_router_probs A__ = router_z_loss_coef A__ = router_aux_loss_coef A__ = self.feed_forward_proj.split("-" ) A__ = act_info[-1] A__ = act_info[0] == "gated" if len(lowercase_ ) > 1 and act_info[0] != "gated" or len(lowercase_ ) > 2: raise ValueError( F"`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer." "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "'gated-gelu' or 'relu'" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": A__ = "gelu_new" super().__init__( pad_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , **lowercase_ , )
355
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : Union[str, Any] =logging.get_logger(__name__) __lowerCAmelCase : Optional[int] =[ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def UpperCamelCase ( _lowerCamelCase : int ): if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: A__ = k.replace(_lowerCamelCase , _lowerCamelCase ) if k.startswith("encoder" ): A__ = k.replace(".attn" , ".self_attn" ) A__ = k.replace("norm1" , "self_attn_layer_norm" ) A__ = k.replace("norm2" , "final_layer_norm" ) elif k.startswith("decoder" ): A__ = k.replace("norm1" , "self_attn_layer_norm" ) A__ = k.replace("norm2" , "encoder_attn_layer_norm" ) A__ = k.replace("norm3" , "final_layer_norm" ) return k def UpperCamelCase ( _lowerCamelCase : int ): A__ = [ "model.encoder.layernorm_embedding.weight", "model.encoder.layernorm_embedding.bias", "model.decoder.layernorm_embedding.weight", "model.decoder.layernorm_embedding.bias", ] for k in keys: A__ = sd.pop(_lowerCamelCase ) A__ = k.replace("layernorm_embedding" , "layer_norm" ) assert new_k not in sd A__ = v __lowerCAmelCase : Optional[int] =["START"] @torch.no_grad() def UpperCamelCase ( _lowerCamelCase : Dict , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any] ): A__ = torch.load(_lowerCamelCase , map_location="cpu" ) A__ = model["model"] A__ = BlenderbotConfig.from_json_file(_lowerCamelCase ) A__ = BlenderbotForConditionalGeneration(_lowerCamelCase ) A__ = m.model.state_dict().keys() A__ = [] A__ = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue A__ = rename_state_dict_key(_lowerCamelCase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: A__ = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(_lowerCamelCase ) m.model.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) m.half() m.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) __lowerCAmelCase : Union[str, Any] =parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
123
0
"""simple docstring""" import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin A_ : Optional[Any] = get_tests_dir("fixtures/test_sentencepiece_bpe.model") class lowerCamelCase (A__ ,unittest.TestCase ): lowerCamelCase__ : List[str] = BartphoTokenizer lowerCamelCase__ : Optional[int] = False lowerCamelCase__ : Optional[Any] = True def SCREAMING_SNAKE_CASE ( self : Dict ) -> Any: super().setUp() SCREAMING_SNAKE_CASE__ = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] SCREAMING_SNAKE_CASE__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) SCREAMING_SNAKE_CASE__ = {"""unk_token""": """<unk>"""} SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F"""{token} {vocab_tokens[token]}\n""" ) SCREAMING_SNAKE_CASE__ = BartphoTokenizer(__UpperCAmelCase , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self : List[Any] , **__UpperCAmelCase : List[str] ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **__UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[Any]: SCREAMING_SNAKE_CASE__ = """This is a là test""" SCREAMING_SNAKE_CASE__ = """This is a<unk><unk> test""" return input_text, output_text def SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: SCREAMING_SNAKE_CASE__ = BartphoTokenizer(__UpperCAmelCase , self.monolingual_vocab_file , **self.special_tokens_map ) SCREAMING_SNAKE_CASE__ = """This is a là test""" SCREAMING_SNAKE_CASE__ = """▁This ▁is ▁a ▁l à ▁t est""".split() SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) SCREAMING_SNAKE_CASE__ = tokens + [tokenizer.unk_token] SCREAMING_SNAKE_CASE__ = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase )
165
'''simple docstring''' import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder a : List[Any] = """__DUMMY_TRANSFORMERS_USER__""" a : Tuple = """Dummy User""" a : Optional[Any] = """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""" a : Optional[Any] = """https://hub-ci.huggingface.co""" a : List[Any] = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" a : Tuple = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" a : str = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def __lowerCamelCase ( _lowercase ) -> Optional[int]: monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""" , _lowercase ) @pytest.fixture def __lowerCamelCase ( _lowercase ) -> List[str]: monkeypatch.setattr("""datasets.config.HF_ENDPOINT""" , _lowercase ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""" , _lowercase ) @pytest.fixture def __lowerCamelCase ( _lowercase ) -> Any: monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""" , _lowercase ) @pytest.fixture def __lowerCamelCase ( _lowercase , _lowercase ) -> Optional[Any]: HfFolder.save_token(_lowercase ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( ) -> str: return HfApi(endpoint=_lowercase ) @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( _lowercase ) -> Union[str, Any]: UpperCAmelCase : str = HfFolder.get_token() HfFolder.save_token(_lowercase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_lowercase ) @pytest.fixture def __lowerCamelCase ( _lowercase ) -> Any: def _cleanup_repo(_lowercase ): hf_api.delete_repo(_lowercase , token=_lowercase , repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def __lowerCamelCase ( _lowercase ) -> List[str]: @contextmanager def _temporary_repo(_lowercase ): try: yield repo_id finally: cleanup_repo(_lowercase ) return _temporary_repo @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> List[str]: UpperCAmelCase : str = F'''repo_txt_data-{int(time.time() * 10e3 )}''' UpperCAmelCase : List[Any] = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_lowercase , token=_lowercase , repo_type="""dataset""" , private=_lowercase ) hf_api.upload_file( token=_lowercase , path_or_fileobj=str(_lowercase ) , path_in_repo="""data/text_data.txt""" , repo_id=_lowercase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_lowercase , token=_lowercase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> List[Any]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> Optional[int]: UpperCAmelCase : Optional[int] = F'''repo_zipped_txt_data-{int(time.time() * 10e3 )}''' UpperCAmelCase : Optional[int] = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_lowercase , token=_lowercase , repo_type="""dataset""" , private=_lowercase ) hf_api.upload_file( token=_lowercase , path_or_fileobj=str(_lowercase ) , path_in_repo="""data.zip""" , repo_id=_lowercase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_lowercase , token=_lowercase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> List[str]: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> Tuple: UpperCAmelCase : List[Any] = F'''repo_zipped_img_data-{int(time.time() * 10e3 )}''' UpperCAmelCase : List[str] = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_lowercase , token=_lowercase , repo_type="""dataset""" , private=_lowercase ) hf_api.upload_file( token=_lowercase , path_or_fileobj=str(_lowercase ) , path_in_repo="""data.zip""" , repo_id=_lowercase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_lowercase , token=_lowercase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> List[Any]: return hf_private_dataset_repo_zipped_img_data_
265
0
from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
359
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def _a ( lowerCamelCase: List[Any] ) -> List[Any]: # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def _a ( ) -> Dict: '''simple docstring''' with parallel_backend('''spark''' ): assert ParallelBackendConfig.backend_name == "spark" __A = [1, 2, 3] with pytest.raises(lowerCamelCase ): with parallel_backend('''unsupported backend''' ): map_nested(lowerCamelCase , lowerCamelCase , num_proc=2 ) with pytest.raises(lowerCamelCase ): with parallel_backend('''unsupported backend''' ): map_nested(lowerCamelCase , lowerCamelCase , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('''num_proc''' , [2, -1] ) def _a ( lowerCamelCase: Optional[int] ) -> List[str]: '''simple docstring''' __A = [1, 2] __A = {'''a''': 1, '''b''': 2} __A = {'''a''': [1, 2], '''b''': [3, 4]} __A = {'''a''': {'''1''': 1}, '''b''': 2} __A = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} __A = [2, 3] __A = {'''a''': 2, '''b''': 3} __A = {'''a''': [2, 3], '''b''': [4, 5]} __A = {'''a''': {'''1''': 2}, '''b''': 3} __A = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} with parallel_backend('''spark''' ): assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa
250
0
'''simple docstring''' from typing import List import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: lowerCamelCase__ : Any = {key: len(UpperCamelCase ) for key, value in gen_kwargs.items() if isinstance(UpperCamelCase , UpperCamelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( """Sharding is ambiguous for this dataset: """ + """we found several data sources lists of different lengths, and we don't know over which list we should parallelize:\n""" + """\n""".join(f'''\t- key {key} has length {length}''' for key, length in lists_lengths.items() ) + """\nTo fix this, check the 'gen_kwargs' and make sure to use lists only for data sources, """ + """and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.""" ) ) lowerCamelCase__ : Any = max(lists_lengths.values() , default=0 ) return max(1 , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[range]: lowerCamelCase__ : Optional[Any] = [] for group_idx in range(UpperCamelCase ): lowerCamelCase__ : Tuple = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break lowerCamelCase__ : List[str] = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 lowerCamelCase__ : Tuple = range(UpperCamelCase , start + num_shards_to_add ) shards_indices_per_group.append(UpperCamelCase ) return shards_indices_per_group def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[dict]: lowerCamelCase__ : Optional[int] = _number_of_shards_in_gen_kwargs(UpperCamelCase ) if num_shards == 1: return [dict(UpperCamelCase )] else: lowerCamelCase__ : Union[str, Any] = _distribute_shards(num_shards=UpperCamelCase , max_num_jobs=UpperCamelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(UpperCamelCase , UpperCamelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(UpperCamelCase ) ) ] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> dict: return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , UpperCamelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> dict: lowerCamelCase__ : int = {len(UpperCamelCase ) for value in gen_kwargs.values() if isinstance(UpperCamelCase , UpperCamelCase )} lowerCamelCase__ : Dict = {} for size in list_sizes: lowerCamelCase__ : int = list(range(UpperCamelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes lowerCamelCase__ : Union[str, Any] = dict(UpperCamelCase ) for key, value in shuffled_kwargs.items(): if isinstance(UpperCamelCase , UpperCamelCase ): lowerCamelCase__ : Dict = [value[i] for i in indices_per_size[len(UpperCamelCase )]] return shuffled_kwargs
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
from PIL import Image def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): def brightness(SCREAMING_SNAKE_CASE_ ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError("level must be between -255.0 (black) and 255.0 (white)" ) return img.point(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": # Load image with Image.open("""image_data/lena.jpg""") as img: # Change brightness to 100 lowercase_ = change_brightness(img, 100) brigt_img.save("""image_data/lena_brightness.png""", format="""png""")
357
import os from collections.abc import Iterator def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = "." ): for dir_path, dir_names, filenames in os.walk(SCREAMING_SNAKE_CASE_ ): lowercase__ = [d for d in dir_names if d != "scripts" and d[0] not in "._"] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(SCREAMING_SNAKE_CASE_ )[1] in (".py", ".ipynb"): yield os.path.join(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ).lstrip("./" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return f'''{i * " "}*''' if i else "\n##" def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(SCREAMING_SNAKE_CASE_ ) or old_parts[i] != new_part) and new_part: print(f'''{md_prefix(SCREAMING_SNAKE_CASE_ )} {new_part.replace("_" , " " ).title()}''' ) return new_path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = "." ): lowercase__ = "" for filepath in sorted(good_file_paths(SCREAMING_SNAKE_CASE_ ) ): lowercase__ , lowercase__ = os.path.split(SCREAMING_SNAKE_CASE_ ) if filepath != old_path: lowercase__ = print_path(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = (filepath.count(os.sep ) + 1) if filepath else 0 lowercase__ = f'''{filepath}/{filename}'''.replace(" " , "%20" ) lowercase__ = os.path.splitext(filename.replace("_" , " " ).title() )[0] print(f'''{md_prefix(SCREAMING_SNAKE_CASE_ )} [{filename}]({url})''' ) if __name__ == "__main__": print_directory_md(""".""")
224
0
'''simple docstring''' from collections.abc import Sequence def _a( UpperCamelCase__ : Sequence[float], UpperCamelCase__ : bool = False ): '''simple docstring''' if not arr: return 0 SCREAMING_SNAKE_CASE__ : int =0 if allow_empty_subarrays else float('''-inf''' ) SCREAMING_SNAKE_CASE__ : Dict =0.0 for num in arr: SCREAMING_SNAKE_CASE__ : Optional[int] =max(0 if allow_empty_subarrays else num, curr_sum + num ) SCREAMING_SNAKE_CASE__ : Optional[Any] =max(UpperCamelCase__, UpperCamelCase__ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() a_ = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'''{max_subarray_sum(nums) = }''')
152
'''simple docstring''' import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_torchaudio, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): snake_case_ = MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING snake_case_ = TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING def __magic_name__ ( self : str , __lowercase : str , __lowercase : List[str] , __lowercase : Union[str, Any] ) -> Dict: SCREAMING_SNAKE_CASE__ : Union[str, Any] =AudioClassificationPipeline(model=__lowercase , feature_extractor=__lowercase ) # test with a raw waveform SCREAMING_SNAKE_CASE__ : Optional[int] =np.zeros((3_40_00,) ) SCREAMING_SNAKE_CASE__ : str =np.zeros((1_40_00,) ) return audio_classifier, [audioa, audio] def __magic_name__ ( self : Optional[int] , __lowercase : int , __lowercase : Optional[int] ) -> str: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Dict =examples SCREAMING_SNAKE_CASE__ : str =audio_classifier(__lowercase ) # by default a model is initialized with num_labels=2 self.assertEqual( __lowercase , [ {'''score''': ANY(__lowercase ), '''label''': ANY(__lowercase )}, {'''score''': ANY(__lowercase ), '''label''': ANY(__lowercase )}, ] , ) SCREAMING_SNAKE_CASE__ : List[Any] =audio_classifier(__lowercase , top_k=1 ) self.assertEqual( __lowercase , [ {'''score''': ANY(__lowercase ), '''label''': ANY(__lowercase )}, ] , ) self.run_torchaudio(__lowercase ) @require_torchaudio def __magic_name__ ( self : Union[str, Any] , __lowercase : str ) -> Optional[Any]: import datasets # test with a local file SCREAMING_SNAKE_CASE__ : Optional[int] =datasets.load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) SCREAMING_SNAKE_CASE__ : int =dataset[0]['''audio''']['''array'''] SCREAMING_SNAKE_CASE__ : Optional[Any] =audio_classifier(__lowercase ) self.assertEqual( __lowercase , [ {'''score''': ANY(__lowercase ), '''label''': ANY(__lowercase )}, {'''score''': ANY(__lowercase ), '''label''': ANY(__lowercase )}, ] , ) @require_torch def __magic_name__ ( self : List[str] ) -> Dict: SCREAMING_SNAKE_CASE__ : Dict ='''anton-l/wav2vec2-random-tiny-classifier''' SCREAMING_SNAKE_CASE__ : Optional[Any] =pipeline('''audio-classification''' , model=__lowercase ) SCREAMING_SNAKE_CASE__ : str =np.ones((80_00,) ) SCREAMING_SNAKE_CASE__ : List[str] =audio_classifier(__lowercase , top_k=4 ) SCREAMING_SNAKE_CASE__ : Dict =[ {'''score''': 0.0842, '''label''': '''no'''}, {'''score''': 0.0838, '''label''': '''up'''}, {'''score''': 0.0837, '''label''': '''go'''}, {'''score''': 0.0834, '''label''': '''right'''}, ] SCREAMING_SNAKE_CASE__ : List[str] =[ {'''score''': 0.0845, '''label''': '''stop'''}, {'''score''': 0.0844, '''label''': '''on'''}, {'''score''': 0.0841, '''label''': '''right'''}, {'''score''': 0.0834, '''label''': '''left'''}, ] self.assertIn(nested_simplify(__lowercase , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) SCREAMING_SNAKE_CASE__ : List[str] ={'''array''': np.ones((80_00,) ), '''sampling_rate''': audio_classifier.feature_extractor.sampling_rate} SCREAMING_SNAKE_CASE__ : Tuple =audio_classifier(__lowercase , top_k=4 ) self.assertIn(nested_simplify(__lowercase , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) @require_torch @slow def __magic_name__ ( self : Dict ) -> Any: import datasets SCREAMING_SNAKE_CASE__ : Union[str, Any] ='''superb/wav2vec2-base-superb-ks''' SCREAMING_SNAKE_CASE__ : Optional[int] =pipeline('''audio-classification''' , model=__lowercase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =datasets.load_dataset('''anton-l/superb_dummy''' , '''ks''' , split='''test''' ) SCREAMING_SNAKE_CASE__ : List[str] =np.array(dataset[3]['''speech'''] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ : int =audio_classifier(__lowercase , top_k=4 ) self.assertEqual( nested_simplify(__lowercase , decimals=3 ) , [ {'''score''': 0.981, '''label''': '''go'''}, {'''score''': 0.007, '''label''': '''up'''}, {'''score''': 0.006, '''label''': '''_unknown_'''}, {'''score''': 0.001, '''label''': '''down'''}, ] , ) @require_tf @unittest.skip('''Audio classification is not implemented for TF''' ) def __magic_name__ ( self : List[str] ) -> Optional[int]: pass
152
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = "▁" _snake_case = {"vocab_file": "sentencepiece.bpe.model"} _snake_case = { "vocab_file": { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model" ), } } _snake_case = { "xlm-roberta-base": 512, "xlm-roberta-large": 512, "xlm-roberta-large-finetuned-conll02-dutch": 512, "xlm-roberta-large-finetuned-conll02-spanish": 512, "xlm-roberta-large-finetuned-conll03-english": 512, "xlm-roberta-large-finetuned-conll03-german": 512, } class lowercase ( UpperCamelCase__ ): _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = ["input_ids", "attention_mask"] def __init__( self , _a , _a="<s>" , _a="</s>" , _a="</s>" , _a="<s>" , _a="<unk>" , _a="<pad>" , _a="<mask>" , _a = None , **_a , ) -> None: # Mask token behave like a normal word, i.e. include the space before it _A : Dict = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else mask_token _A : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) _A : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) _A : Optional[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _A : Tuple = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _A : Any = 1 _A : List[Any] = len(self.sp_model ) + self.fairseq_offset _A : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> Any: _A : str = self.__dict__.copy() _A : int = None _A : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self , _a ) -> int: _A : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _A : List[str] = {} _A : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def a__ ( self , _a , _a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _A : Optional[Any] = [self.cls_token_id] _A : List[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def a__ ( self , _a , _a = None , _a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def a__ ( self , _a , _a = None ) -> List[int]: _A : Optional[Any] = [self.sep_token_id] _A : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def a__ ( self ) -> Dict: return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def a__ ( self ) -> str: _A : Dict = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a__ ( self , _a ) -> List[str]: return self.sp_model.encode(_a , out_type=_a ) def a__ ( self , _a ) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _A : int = self.sp_model.PieceToId(_a ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a__ ( self , _a ) -> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def a__ ( self , _a ) -> Dict: _A : int = """""".join(_a ).replace(_a , """ """ ).strip() return out_string def a__ ( self , _a , _a = None ) -> Tuple[str]: if not os.path.isdir(_a ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A : str = os.path.join( _a , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , """wb""" ) as fi: _A : Tuple = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
343
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _snake_case = "https://www.indeed.co.in/jobs?q=mobile+app+development&l=" def lowerCAmelCase_ ( snake_case_ = "mumbai" ): _A : Optional[Any] = BeautifulSoup(requests.get(url + location ).content,"""html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""",attrs={"""data-tn-component""": """organicJob"""} ): _A : Tuple = job.find("""a""",attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() _A : Optional[int] = job.find("""span""",{"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("Bangalore"), 1): print(f"""Job {i:>2} is {job[0]} at {job[1]}""")
343
1
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class _snake_case : def __init__( self , a__ , a__=3 , a__=7 , a__=True , a__=True , a__=False , a__=True , a__=99 , a__=32 , a__=5 , a__=4 , a__=37 , a__="gelu" , a__=0.1 , a__=0.1 , a__=512 , a__=16 , a__=2 , a__=0.0_2 , a__=3 , a__=4 , a__=None , ) -> List[str]: '''simple docstring''' snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a__ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a__ , ) def lowerCAmelCase__ ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ) -> Optional[Any]: '''simple docstring''' snake_case_ = FalconModel(config=a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , attention_mask=a__ ) snake_case_ = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ) -> Optional[int]: '''simple docstring''' snake_case_ = True snake_case_ = FalconModel(a__ ) model.to(a__ ) model.eval() snake_case_ = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , ) snake_case_ = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , ) snake_case_ = model(a__ , attention_mask=a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ) -> str: '''simple docstring''' snake_case_ = FalconForCausalLM(config=a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , attention_mask=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ) -> Dict: '''simple docstring''' snake_case_ = True snake_case_ = True snake_case_ = FalconForCausalLM(config=a__ ) model.to(a__ ) model.eval() # first forward pass snake_case_ = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , use_cache=a__ , ) snake_case_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case_ = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case_ = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , output_hidden_states=a__ , )["hidden_states"][0] snake_case_ = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , past_key_values=a__ , output_hidden_states=a__ , )["hidden_states"][0] # select random slice snake_case_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a__ , a__ , atol=1e-3 ) ) def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _snake_case ( lowercase_ , lowercase_ , lowercase_ , unittest.TestCase ): lowerCAmelCase_ : Optional[int] = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) lowerCAmelCase_ : List[Any] = (FalconForCausalLM,) if is_torch_available() else () lowerCAmelCase_ : Union[str, Any] = ( { "feature-extraction": FalconModel, "text-classification": FalconForSequenceClassification, "text-generation": FalconForCausalLM, "question-answering": FalconForQuestionAnswering, "token-classification": FalconForTokenClassification, "zero-shot": FalconForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ : Optional[Any] = False lowerCAmelCase_ : List[Any] = False def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ = FalconModelTester(self ) snake_case_ = ConfigTester(self , config_class=a__ , hidden_size=37 ) def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def lowerCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ , *snake_case_ = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: snake_case_ = alibi self.model_tester.create_and_check_model(a__ , *a__ ) def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(a__ ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = FalconForSequenceClassification(a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = "single_label_classification" snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(a__ ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = FalconForSequenceClassification(a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self ) -> str: '''simple docstring''' snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = input_dict["input_ids"] snake_case_ = FalconForCausalLM(a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , use_cache=a__ ) snake_case_ = input_ids.shape[0] snake_case_ = model._convert_to_rw_cache(result.past_key_values ) snake_case_ = model._convert_cache_to_standard_format(a__ , a__ ) for layer in range(len(a__ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = "multi_label_classification" snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(a__ ) snake_case_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) snake_case_ = FalconForSequenceClassification(a__ ) model.to(a__ ) model.eval() snake_case_ = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self ) -> str: '''simple docstring''' for model_class in self.all_generative_model_classes: snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a__ , "use_cache" ): return snake_case_ = model_class(a__ ).to(a__ ) if "use_cache" not in inputs: snake_case_ = True snake_case_ = model(**a__ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return snake_case_ = ( getattr(a__ , "decoder_layers" , a__ ) or getattr(a__ , "num_decoder_layers" , a__ ) or config.num_hidden_layers ) snake_case_ = getattr(a__ , "num_kv_heads" , config.num_attention_heads ) snake_case_ = getattr(a__ , "d_model" , config.hidden_size ) snake_case_ = embed_dim // num_attention_heads snake_case_ = outputs["past_key_values"] self.assertEqual(len(a__ ) , a__ ) snake_case_ , snake_case_ = inputs["input_ids"].shape for i in range(a__ ): if config.new_decoder_architecture: snake_case_ = config.num_attention_heads elif config.multi_query: snake_case_ = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class _snake_case ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) snake_case_ = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a__ ) snake_case_ = tokenizer("My favorite food is" , return_tensors="pt" ).to(a__ ) snake_case_ = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) snake_case_ = model.generate(**a__ , do_sample=a__ , max_new_tokens=19 ) snake_case_ = tokenizer.batch_decode(a__ )[0] self.assertEqual(a__ , a__ ) @slow def lowerCAmelCase__ ( self ) -> Any: '''simple docstring''' for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: snake_case_ = AutoTokenizer.from_pretrained(a__ ) snake_case_ = FalconForCausalLM.from_pretrained(a__ ) model.eval() model.to(a__ ) snake_case_ = tokenizer("My favorite food is" , return_tensors="pt" ).to(a__ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a__ , do_sample=a__ , max_new_tokens=4 ) model.generate(**a__ , do_sample=a__ , max_new_tokens=4 ) model.generate(**a__ , num_beams=2 , max_new_tokens=4 ) @slow def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: snake_case_ = AutoTokenizer.from_pretrained(a__ ) snake_case_ = FalconForCausalLM.from_pretrained(a__ ) model.eval() model.to(device=a__ ) snake_case_ = tokenizer("My favorite food is" , return_tensors="pt" ).to(a__ ) # Test results are the same with and without cache snake_case_ = model.generate(**a__ , do_sample=a__ , max_new_tokens=20 , use_cache=a__ ) snake_case_ = model.generate(**a__ , do_sample=a__ , max_new_tokens=20 , use_cache=a__ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
85
'''simple docstring''' import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _SCREAMING_SNAKE_CASE : Any = False try: _SCREAMING_SNAKE_CASE : Optional[Any] = _is_package_available("google.colab") except ModuleNotFoundError: pass @input.register class _snake_case : def __init__( self , a__ = None , a__ = [] ) -> List[str]: '''simple docstring''' snake_case_ = 0 snake_case_ = choices snake_case_ = prompt if sys.platform == "win32": snake_case_ = "*" else: snake_case_ = "➔ " def lowerCAmelCase__ ( self , a__ , a__ = "" ) -> int: '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 32 , a__ ) else: forceWrite(self.choices[index] , a__ ) def lowerCAmelCase__ ( self , a__ ) -> Tuple: '''simple docstring''' if index == self.position: forceWrite(F' {self.arrow_char} ' ) self.write_choice(a__ ) else: forceWrite(F' {self.choices[index]}' ) reset_cursor() def lowerCAmelCase__ ( self , a__ , a__ = 1 ) -> List[str]: '''simple docstring''' snake_case_ = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(a__ ) move_cursor(a__ , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["up"] ) def lowerCAmelCase__ ( self ) -> Dict: '''simple docstring''' self.move_direction(Direction.UP ) @input.mark(KEYMAP["down"] ) def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["newline"] ) def lowerCAmelCase__ ( self ) -> str: '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) return self.position @input.mark(KEYMAP["interrupt"] ) def lowerCAmelCase__ ( self ) -> Tuple: '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(a__ )] for number in range(10 )] ) def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ = int(chr(self.current_selection ) ) snake_case_ = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , a__ ) else: return else: return def lowerCAmelCase__ ( self , a__ = 0 ) -> List[str]: '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , "\n" ) if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n" ) else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n" ) snake_case_ = default_choice for i in range(len(self.choices ) ): self.print_choice(a__ ) forceWrite("\n" ) move_cursor(len(self.choices ) - self.position , "UP" ) with cursor.hide(): while True: if in_colab: try: snake_case_ = int(builtins.input() ) except ValueError: snake_case_ = default_choice else: snake_case_ = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , "UP" ) clear_line() self.write_choice(a__ , "\n" ) return choice
85
1
def lowerCAmelCase_ (lowerCAmelCase__: int = 6_0_0_8_5_1_4_7_5_1_4_3 ): """simple docstring""" try: UpperCAmelCase_: Any = int(lowerCAmelCase__ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) UpperCAmelCase_: Dict = 2 UpperCAmelCase_: List[str] = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 UpperCAmelCase_: Optional[int] = i while n % i == 0: UpperCAmelCase_: Dict = n // i i += 1 return int(lowerCAmelCase__ ) if __name__ == "__main__": print(F'''{solution() = }''')
350
a : Tuple = 'Tobias Carryer' from time import time class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=int(time() ) ) -> List[Any]: # noqa: B008 UpperCAmelCase_: List[str] = multiplier UpperCAmelCase_: Tuple = increment UpperCAmelCase_: Tuple = modulo UpperCAmelCase_: List[str] = seed def __snake_case (self ) -> Any: UpperCAmelCase_: List[str] = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. a : Optional[int] = LinearCongruentialGenerator(1_664_525, 1_013_904_223, 2 << 31) while True: print(lcg.next_number())
82
0
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer _lowerCamelCase : str = ["""gpt2"""] _lowerCamelCase : str = """gpt2""" if is_tf_available(): class UpperCamelCase_ ( tf.Module ): '''simple docstring''' def __init__( self : Tuple , UpperCAmelCase__ : Union[str, Any]) ->List[Any]: '''simple docstring''' super().__init__() A__ = tokenizer A__ = AutoConfig.from_pretrained(UpperCAmelCase__) A__ = TFGPTaLMHeadModel.from_config(UpperCAmelCase__) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name='''text'''),)) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : List[str]) ->Optional[int]: '''simple docstring''' A__ = self.tokenizer(UpperCAmelCase__) A__ = tokenized['''input_ids'''].to_tensor() A__ = tf.cast(input_ids_dense > 0 , tf.intaa) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) A__ = self.model(input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__)['''logits'''] return outputs @require_tf @require_keras_nlp class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Dict: '''simple docstring''' super().setUp() A__ = [GPTaTokenizer.from_pretrained(UpperCAmelCase__) for checkpoint in (TOKENIZER_CHECKPOINTS)] A__ = [TFGPTaTokenizer.from_pretrained(UpperCAmelCase__) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers) == len(self.tf_tokenizers) A__ = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] A__ = list(zip(self.test_sentences , self.test_sentences[::-1])) def SCREAMING_SNAKE_CASE ( self : List[str]) ->Dict: '''simple docstring''' for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers): for test_inputs in self.test_sentences: A__ = tokenizer([test_inputs] , return_tensors='''tf''') A__ = tf_tokenizer([test_inputs]) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors A__ = python_outputs[key].numpy() A__ = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape)) self.assertTrue(tf.reduce_all(tf.cast(UpperCAmelCase__ , tf.intaa) == tf_outputs_values)) @slow def SCREAMING_SNAKE_CASE ( self : int) ->int: '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: A__ = tf.function(UpperCAmelCase__) for test_inputs in self.test_sentences: A__ = tf.constant(UpperCAmelCase__) A__ = compiled_tokenizer(UpperCAmelCase__) A__ = tf_tokenizer(UpperCAmelCase__) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key])) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Optional[Any]: '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: A__ = ModelToSave(tokenizer=UpperCAmelCase__) A__ = tf.convert_to_tensor([self.test_sentences[0]]) A__ = model.serving(UpperCAmelCase__) # Build model with some sample inputs with TemporaryDirectory() as tempdir: A__ = Path(UpperCAmelCase__) / '''saved.model''' tf.saved_model.save(UpperCAmelCase__ , UpperCAmelCase__ , signatures={'''serving_default''': model.serving}) A__ = tf.saved_model.load(UpperCAmelCase__) A__ = loaded_model.signatures['''serving_default'''](UpperCAmelCase__)['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output)) @slow def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: A__ = tf.convert_to_tensor([self.test_sentences[0]]) A__ = tf_tokenizer(UpperCAmelCase__) # Build model with some sample inputs A__ = tf_tokenizer.get_config() A__ = TFGPTaTokenizer.from_config(UpperCAmelCase__) A__ = model_from_config(UpperCAmelCase__) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key])) @slow def SCREAMING_SNAKE_CASE ( self : Dict) ->List[Any]: '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: # for the test to run A__ = 123_123 for max_length in [3, 5, 1_024]: A__ = tf.convert_to_tensor([self.test_sentences[0]]) A__ = tf_tokenizer(UpperCAmelCase__ , max_length=UpperCAmelCase__) A__ = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
14
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np _lowerCamelCase : str = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 _lowerCamelCase : Tuple = typing.Union[np.floataa, int, float] # noqa: UP007 def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> VectorOut: """simple docstring""" return np.sqrt(np.sum((np.asarray(lowercase_ ) - np.asarray(lowercase_ )) ** 2 ) ) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> VectorOut: """simple docstring""" return sum((va - va) ** 2 for va, va in zip(lowercase_ , lowercase_ ) ) ** (1 / 2) if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( ) -> None: """simple docstring""" from timeit import timeit print('''Without Numpy''' ) print( timeit( '''euclidean_distance_no_np([1, 2, 3], [4, 5, 6])''' , number=10_000 , globals=globals() , ) ) print('''With Numpy''' ) print( timeit( '''euclidean_distance([1, 2, 3], [4, 5, 6])''' , number=10_000 , globals=globals() , ) ) benchmark()
14
1
"""simple docstring""" import re import string import numpy as np import datasets __A = "\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n" __A = "\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results[\"exact_match\"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"]\n >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 33.3\n\n" __A = "\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): def lowercase_ ( self : List[Any])-> Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence"), "references": datasets.Value("string" , id="sequence"), }) , reference_urls=[] , ) def lowercase_ ( self : str , UpperCamelCase__ : str , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : int=False , )-> Union[str, Any]: '''simple docstring''' if regexes_to_ignore is not None: for s in regexes_to_ignore: __lowerCAmelCase: Optional[int] = np.array([re.sub(UpperCamelCase__ , "" , UpperCamelCase__) for x in predictions]) __lowerCAmelCase: Optional[int] = np.array([re.sub(UpperCamelCase__ , "" , UpperCamelCase__) for x in references]) else: __lowerCAmelCase: int = np.asarray(UpperCamelCase__) __lowerCAmelCase: Optional[int] = np.asarray(UpperCamelCase__) if ignore_case: __lowerCAmelCase: int = np.char.lower(UpperCamelCase__) __lowerCAmelCase: str = np.char.lower(UpperCamelCase__) if ignore_punctuation: __lowerCAmelCase: Optional[int] = string.punctuation.maketrans("" , "" , string.punctuation) __lowerCAmelCase: Any = np.char.translate(UpperCamelCase__ , table=UpperCamelCase__) __lowerCAmelCase: Dict = np.char.translate(UpperCamelCase__ , table=UpperCamelCase__) if ignore_numbers: __lowerCAmelCase: Union[str, Any] = string.digits.maketrans("" , "" , string.digits) __lowerCAmelCase: Optional[int] = np.char.translate(UpperCamelCase__ , table=UpperCamelCase__) __lowerCAmelCase: Optional[int] = np.char.translate(UpperCamelCase__ , table=UpperCamelCase__) __lowerCAmelCase: Dict = predictions == references return {"exact_match": np.mean(UpperCamelCase__) * 1_0_0}
108
"""simple docstring""" from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def a__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.ndarray: __lowerCAmelCase: List[Any] = cva.getAffineTransform(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return cva.warpAffine(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (rows, cols) ) if __name__ == "__main__": # read original image __A = cva.imread( str(Path(__file__).resolve().parent.parent / "image_data" / "lena.jpg") ) # turn image in gray scale value __A = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape __A , __A = gray_img.shape # set different points to rotate image __A = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) __A = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) __A = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) __A = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list __A = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations __A = plt.figure(1) __A = ["Original", "Rotation 1", "Rotation 2", "Rotation 3"] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, "gray") plt.title(titles[i]) plt.axis("off") plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
108
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: snake_case_ : Union[str, Any] = None snake_case_ : Union[str, Any] = logging.get_logger(__name__) snake_case_ : List[str] = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} snake_case_ : Tuple = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } snake_case_ : Tuple = { "google/bigbird-roberta-base": 4096, "google/bigbird-roberta-large": 4096, "google/bigbird-base-trivia-itc": 4096, } snake_case_ : Dict = "▁" class __snake_case ( a ): UpperCAmelCase__ : Optional[Any] = VOCAB_FILES_NAMES UpperCAmelCase__ : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : List[str] = BigBirdTokenizer UpperCAmelCase__ : Any = ['''input_ids''', '''attention_mask'''] UpperCAmelCase__ : List[int] = [] def __init__( self : List[str] , _snake_case : Dict=None , _snake_case : List[Any]=None , _snake_case : str="<unk>" , _snake_case : Optional[int]="<s>" , _snake_case : Any="</s>" , _snake_case : List[Any]="<pad>" , _snake_case : Tuple="[SEP]" , _snake_case : List[Any]="[MASK]" , _snake_case : str="[CLS]" , **_snake_case : List[Any] , ): """simple docstring""" UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else bos_token UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else eos_token UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else unk_token UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else pad_token UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else cls_token UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else mask_token super().__init__( _snake_case , tokenizer_file=_snake_case , bos_token=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , **_snake_case , ) UpperCAmelCase_ = vocab_file UpperCAmelCase_ = False if not self.vocab_file else True def lowerCamelCase ( self : Any , _snake_case : List[int] , _snake_case : Optional[List[int]] = None): """simple docstring""" UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self : Optional[Any] , _snake_case : List[int] , _snake_case : Optional[List[int]] = None , _snake_case : bool = False): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''') return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(_snake_case)) + [1] return [1] + ([0] * len(_snake_case)) + [1] + ([0] * len(_snake_case)) + [1] def lowerCamelCase ( self : str , _snake_case : List[int] , _snake_case : Optional[List[int]] = None): """simple docstring""" UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def lowerCamelCase ( self : Tuple , _snake_case : str , _snake_case : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''') if not os.path.isdir(_snake_case): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""") return UpperCAmelCase_ = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(_snake_case): copyfile(self.vocab_file , _snake_case) return (out_vocab_file,)
51
from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class __snake_case : pass
51
1
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' if upper_limit < 0: raise ValueError("""Limit for the Catalan sequence must be ≥ 0""" ) UpperCAmelCase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 UpperCAmelCase__ = 1 if upper_limit > 0: UpperCAmelCase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE__ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: UpperCAmelCase_ = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(f"The Catalan numbers from 0 through {N} are:") print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
366
'''simple docstring''' import enum import shutil import sys UpperCAmelCase_ , UpperCAmelCase_ = shutil.get_terminal_size() UpperCAmelCase_ = {'UP': 'A', 'DOWN': 'B', 'RIGHT': 'C', 'LEFT': 'D'} class lowerCAmelCase_ ( enum.Enum ): '''simple docstring''' lowerCAmelCase_ : int = 0 lowerCAmelCase_ : Union[str, Any] = 1 def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[Any]="" ): '''simple docstring''' sys.stdout.write(str(SCREAMING_SNAKE_CASE__ ) + end ) sys.stdout.flush() def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : int="" ): '''simple docstring''' forceWrite(F'''\u001b[{color}m{content}\u001b[0m''' , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( ): '''simple docstring''' forceWrite("""\r""" ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' forceWrite(F'''\033[{num_lines}{CURSOR_TO_CHAR[direction.upper()]}''' ) def _UpperCamelCase ( ): '''simple docstring''' forceWrite(""" """ * TERMINAL_WIDTH ) reset_cursor() def _UpperCamelCase ( ): '''simple docstring''' reset_cursor() forceWrite("""-""" * TERMINAL_WIDTH )
61
0
'''simple docstring''' SCREAMING_SNAKE_CASE_: Union[str, Any] ={ 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def lowerCAmelCase_ ( snake_case_ : dict , snake_case_ : Dict , snake_case_ : List[Any] ) -> list[str]: '''simple docstring''' UpperCAmelCase_ = set() # keep track of all the paths to be checked UpperCAmelCase_ = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue UpperCAmelCase_ = queue.pop(0 ) # get the last node from the path UpperCAmelCase_ = path[-1] if node not in explored: UpperCAmelCase_ = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: UpperCAmelCase_ = list(snake_case_ ) new_path.append(snake_case_ ) queue.append(snake_case_ ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(snake_case_ ) # in case there's no path between the 2 nodes return [] def lowerCAmelCase_ ( snake_case_ : dict , snake_case_ : int , snake_case_ : int ) -> int: '''simple docstring''' if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 UpperCAmelCase_ = [start] UpperCAmelCase_ = set(snake_case_ ) # Keep tab on distances from `start` node. UpperCAmelCase_ = {start: 0, target: -1} while queue: UpperCAmelCase_ = queue.pop(0 ) if node == target: UpperCAmelCase_ = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(snake_case_ ) queue.append(snake_case_ ) UpperCAmelCase_ = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
1
'''simple docstring''' from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch SCREAMING_SNAKE_CASE_: Optional[int] =logging.get_logger(__name__) @add_end_docstrings( UpperCamelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class __A ( UpperCamelCase__ ): def _lowercase (self : str , __a : GenericTensor ): if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=__a ) else: raise ValueError("Unsupported framework" ) return masked_index def _lowercase (self : Tuple , __a : GenericTensor ): UpperCAmelCase_ = self.get_masked_index(__a ) UpperCAmelCase_ = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( "fill-mask" , self.model.base_model_prefix , f"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def _lowercase (self : List[Any] , __a : GenericTensor ): if isinstance(__a , __a ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input["input_ids"][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(__a ) def _lowercase (self : Tuple , __a : Dict , __a : List[str]=None , **__a : Any ): if return_tensors is None: UpperCAmelCase_ = self.framework UpperCAmelCase_ = self.tokenizer(__a , return_tensors=__a ) self.ensure_exactly_one_mask_token(__a ) return model_inputs def _lowercase (self : str , __a : Optional[int] ): UpperCAmelCase_ = self.model(**__a ) UpperCAmelCase_ = model_inputs["input_ids"] return model_outputs def _lowercase (self : List[str] , __a : Tuple , __a : int=5 , __a : Dict=None ): # Cap top_k if there are targets if target_ids is not None and target_ids.shape[0] < top_k: UpperCAmelCase_ = target_ids.shape[0] UpperCAmelCase_ = model_outputs["input_ids"][0] UpperCAmelCase_ = model_outputs["logits"] if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] UpperCAmelCase_ = outputs.numpy() UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = stable_softmax(__a , axis=-1 ) if target_ids is not None: UpperCAmelCase_ = tf.gather_nd(tf.squeeze(__a , 0 ) , target_ids.reshape(-1 , 1 ) ) UpperCAmelCase_ = tf.expand_dims(__a , 0 ) UpperCAmelCase_ = tf.math.top_k(__a , k=__a ) UpperCAmelCase_ , UpperCAmelCase_ = topk.values.numpy(), topk.indices.numpy() else: UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=__a ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = logits.softmax(dim=-1 ) if target_ids is not None: UpperCAmelCase_ = probs[..., target_ids] UpperCAmelCase_ , UpperCAmelCase_ = probs.topk(__a ) UpperCAmelCase_ = [] UpperCAmelCase_ = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): UpperCAmelCase_ = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place UpperCAmelCase_ = input_ids.numpy().copy() if target_ids is not None: UpperCAmelCase_ = target_ids[p].tolist() UpperCAmelCase_ = p # Filter padding out: UpperCAmelCase_ = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back UpperCAmelCase_ = self.tokenizer.decode(__a , skip_special_tokens=__a ) UpperCAmelCase_ = {"score": v, "token": p, "token_str": self.tokenizer.decode([p] ), "sequence": sequence} row.append(__a ) result.append(__a ) if single_mask: return result[0] return result def _lowercase (self : Dict , __a : List[Any] , __a : List[str]=None ): if isinstance(__a , __a ): UpperCAmelCase_ = [targets] try: UpperCAmelCase_ = self.tokenizer.get_vocab() except Exception: UpperCAmelCase_ = {} UpperCAmelCase_ = [] for target in targets: UpperCAmelCase_ = vocab.get(__a , __a ) if id_ is None: UpperCAmelCase_ = self.tokenizer( __a , add_special_tokens=__a , return_attention_mask=__a , return_token_type_ids=__a , max_length=1 , truncation=__a , )["input_ids"] if len(__a ) == 0: logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ "We cannot replace it with anything meaningful, ignoring it" ) continue UpperCAmelCase_ = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ f"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) UpperCAmelCase_ = list(set(__a ) ) if len(__a ) == 0: raise ValueError("At least one target must be provided when passed." ) UpperCAmelCase_ = np.array(__a ) return target_ids def _lowercase (self : Tuple , __a : Dict=None , __a : List[str]=None ): UpperCAmelCase_ = {} if targets is not None: UpperCAmelCase_ = self.get_target_ids(__a , __a ) UpperCAmelCase_ = target_ids if top_k is not None: UpperCAmelCase_ = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( "fill-mask" , self.model.base_model_prefix , "The tokenizer does not define a `mask_token`." ) return {}, {}, postprocess_params def __call__(self : Union[str, Any] , __a : str , *__a : Any , **__a : Tuple ): UpperCAmelCase_ = super().__call__(__a , **__a ) if isinstance(__a , __a ) and len(__a ) == 1: return outputs[0] return outputs
1
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging UpperCAmelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' __snake_case = ["pixel_values"] def __init__( self , _UpperCAmelCase = True , _UpperCAmelCase = None , _UpperCAmelCase = PILImageResampling.BILINEAR , _UpperCAmelCase = True , _UpperCAmelCase = None , _UpperCAmelCase = True , _UpperCAmelCase = 1 / 2_55 , _UpperCAmelCase = True , _UpperCAmelCase = None , _UpperCAmelCase = None , **_UpperCAmelCase , ): super().__init__(**_UpperCAmelCase ) snake_case_ = size if size is not None else {'''shortest_edge''': 2_56} snake_case_ = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) snake_case_ = crop_size if crop_size is not None else {'''height''': 2_24, '''width''': 2_24} snake_case_ = get_size_dict(_UpperCAmelCase ) snake_case_ = do_resize snake_case_ = size snake_case_ = resample snake_case_ = do_center_crop snake_case_ = crop_size snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = PILImageResampling.BICUBIC , _UpperCAmelCase = None , **_UpperCAmelCase , ): snake_case_ = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) snake_case_ = get_resize_output_image_size(_UpperCAmelCase , size=size['''shortest_edge'''] , default_to_square=_UpperCAmelCase ) return resize(_UpperCAmelCase , size=_UpperCAmelCase , resample=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , **_UpperCAmelCase , ): snake_case_ = get_size_dict(_UpperCAmelCase ) return center_crop(_UpperCAmelCase , size=(size['''height'''], size['''width''']) , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , **_UpperCAmelCase ): return rescale(_UpperCAmelCase , scale=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , **_UpperCAmelCase , ): return normalize(_UpperCAmelCase , mean=_UpperCAmelCase , std=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = ChannelDimension.FIRST , **_UpperCAmelCase , ): snake_case_ = do_resize if do_resize is not None else self.do_resize snake_case_ = size if size is not None else self.size snake_case_ = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) snake_case_ = resample if resample is not None else self.resample snake_case_ = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ = crop_size if crop_size is not None else self.crop_size snake_case_ = get_size_dict(_UpperCAmelCase ) snake_case_ = do_rescale if do_rescale is not None else self.do_rescale snake_case_ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ = do_normalize if do_normalize is not None else self.do_normalize snake_case_ = image_mean if image_mean is not None else self.image_mean snake_case_ = image_std if image_std is not None else self.image_std snake_case_ = make_list_of_images(_UpperCAmelCase ) if not valid_images(_UpperCAmelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. snake_case_ = [to_numpy_array(_UpperCAmelCase ) for image in images] if do_resize: snake_case_ = [self.resize(image=_UpperCAmelCase , size=_UpperCAmelCase , resample=_UpperCAmelCase ) for image in images] if do_center_crop: snake_case_ = [self.center_crop(image=_UpperCAmelCase , size=_UpperCAmelCase ) for image in images] if do_rescale: snake_case_ = [self.rescale(image=_UpperCAmelCase , scale=_UpperCAmelCase ) for image in images] if do_normalize: snake_case_ = [self.normalize(image=_UpperCAmelCase , mean=_UpperCAmelCase , std=_UpperCAmelCase ) for image in images] snake_case_ = [to_channel_dimension_format(_UpperCAmelCase , _UpperCAmelCase ) for image in images] snake_case_ = {'''pixel_values''': images} return BatchFeature(data=_UpperCAmelCase , tensor_type=_UpperCAmelCase )
267
import numpy as np import datasets UpperCAmelCase = """ Compute the Mahalanobis Distance Mahalonobis distance is the distance between a point and a distribution. And not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance. It was introduced by Prof. P. C. Mahalanobis in 1936 and has been used in various statistical applications ever since [source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/] """ UpperCAmelCase = """\ @article{de2000mahalanobis, title={The mahalanobis distance}, author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L}, journal={Chemometrics and intelligent laboratory systems}, volume={50}, number={1}, pages={1--18}, year={2000}, publisher={Elsevier} } """ UpperCAmelCase = """ Args: X: List of datapoints to be compared with the `reference_distribution`. reference_distribution: List of datapoints from the reference distribution we want to compare to. Returns: mahalanobis: The Mahalonobis distance for each datapoint in `X`. Examples: >>> mahalanobis_metric = datasets.load_metric(\"mahalanobis\") >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]]) >>> print(results) {'mahalanobis': array([0.5])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''X''': datasets.Sequence(datasets.Value('''float''' , id='''sequence''' ) , id='''X''' ), } ) , ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase ): # convert to numpy arrays snake_case_ = np.array(_UpperCAmelCase ) snake_case_ = np.array(_UpperCAmelCase ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError('''Expected `X` to be a 2D vector''' ) if len(reference_distribution.shape ) != 2: raise ValueError('''Expected `reference_distribution` to be a 2D vector''' ) if reference_distribution.shape[0] < 2: raise ValueError( '''Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension''' ) # Get mahalanobis distance for each prediction snake_case_ = X - np.mean(_UpperCAmelCase ) snake_case_ = np.cov(reference_distribution.T ) try: snake_case_ = np.linalg.inv(_UpperCAmelCase ) except np.linalg.LinAlgError: snake_case_ = np.linalg.pinv(_UpperCAmelCase ) snake_case_ = np.dot(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ = np.dot(_UpperCAmelCase , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
267
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType snake_case_ = logging.get_logger(__name__) snake_case_ = { """openai/imagegpt-small""": """""", """openai/imagegpt-medium""": """""", """openai/imagegpt-large""": """""", } class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = """imagegpt""" __UpperCamelCase = ["""past_key_values"""] __UpperCamelCase = { """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self :int , lowercase_ :List[str]=5_12 + 1 , lowercase_ :Dict=32 * 32 , lowercase_ :List[str]=5_12 , lowercase_ :str=24 , lowercase_ :List[Any]=8 , lowercase_ :Optional[int]=None , lowercase_ :List[Any]="quick_gelu" , lowercase_ :str=0.1 , lowercase_ :List[Any]=0.1 , lowercase_ :List[str]=0.1 , lowercase_ :List[str]=1E-5 , lowercase_ :Tuple=0.02 , lowercase_ :List[str]=True , lowercase_ :Optional[int]=True , lowercase_ :str=False , lowercase_ :Union[str, Any]=False , lowercase_ :List[str]=False , **lowercase_ :int , ) -> Optional[Any]: UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = n_inner UpperCAmelCase = activation_function UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = attn_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = scale_attn_weights UpperCAmelCase = use_cache UpperCAmelCase = scale_attn_by_inverse_layer_idx UpperCAmelCase = reorder_and_upcast_attn UpperCAmelCase = tie_word_embeddings super().__init__(tie_word_embeddings=lowercase_ , **lowercase_ ) class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" @property def UpperCAmelCase__ ( self :int ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ] ) def UpperCAmelCase__ ( self :Optional[Any] , lowercase_ :"FeatureExtractionMixin" , lowercase_ :int = 1 , lowercase_ :int = -1 , lowercase_ :bool = False , lowercase_ :Optional["TensorType"] = None , lowercase_ :int = 3 , lowercase_ :int = 32 , lowercase_ :int = 32 , ) -> Mapping[str, Any]: UpperCAmelCase = self._generate_dummy_images(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase = dict(preprocessor(images=lowercase_ , return_tensors=lowercase_ ) ) return inputs
78
"""simple docstring""" class A_ : """simple docstring""" def __init__( self :List[str] , lowercase_ :int , lowercase_ :Optional[int]=None , lowercase_ :List[str]=None ) -> str: UpperCAmelCase = data UpperCAmelCase = previous UpperCAmelCase = next_node def __str__( self :Optional[Any] ) -> str: return f"""{self.data}""" def UpperCAmelCase__ ( self :int ) -> int: return self.data def UpperCAmelCase__ ( self :List[str] ) -> Any: return self.next def UpperCAmelCase__ ( self :Tuple ) -> Optional[int]: return self.previous class A_ : """simple docstring""" def __init__( self :Optional[Any] , lowercase_ :Optional[Any] ) -> str: UpperCAmelCase = head def __iter__( self :List[str] ) -> List[str]: return self def UpperCAmelCase__ ( self :int ) -> Any: if not self.current: raise StopIteration else: UpperCAmelCase = self.current.get_data() UpperCAmelCase = self.current.get_next() return value class A_ : """simple docstring""" def __init__( self :Union[str, Any] ) -> List[Any]: UpperCAmelCase = None # First node in list UpperCAmelCase = None # Last node in list def __str__( self :List[Any] ) -> Optional[Any]: UpperCAmelCase = self.head UpperCAmelCase = [] while current is not None: nodes.append(current.get_data() ) UpperCAmelCase = current.get_next() return " ".join(str(lowercase_ ) for node in nodes ) def __contains__( self :str , lowercase_ :int ) -> str: UpperCAmelCase = self.head while current: if current.get_data() == value: return True UpperCAmelCase = current.get_next() return False def __iter__( self :Tuple ) -> Dict: return LinkedListIterator(self.head ) def UpperCAmelCase__ ( self :Optional[int] ) -> Optional[Any]: if self.head: return self.head.get_data() return None def UpperCAmelCase__ ( self :Union[str, Any] ) -> List[str]: if self.tail: return self.tail.get_data() return None def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :Node ) -> None: if self.head is None: UpperCAmelCase = node UpperCAmelCase = node else: self.insert_before_node(self.head , lowercase_ ) def UpperCAmelCase__ ( self :Any , lowercase_ :Node ) -> None: if self.head is None: self.set_head(lowercase_ ) else: self.insert_after_node(self.tail , lowercase_ ) def UpperCAmelCase__ ( self :List[str] , lowercase_ :int ) -> None: UpperCAmelCase = Node(lowercase_ ) if self.head is None: self.set_head(lowercase_ ) else: self.set_tail(lowercase_ ) def UpperCAmelCase__ ( self :int , lowercase_ :Node , lowercase_ :Node ) -> None: UpperCAmelCase = node UpperCAmelCase = node.previous if node.get_previous() is None: UpperCAmelCase = node_to_insert else: UpperCAmelCase = node_to_insert UpperCAmelCase = node_to_insert def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :Node , lowercase_ :Node ) -> None: UpperCAmelCase = node UpperCAmelCase = node.next if node.get_next() is None: UpperCAmelCase = node_to_insert else: UpperCAmelCase = node_to_insert UpperCAmelCase = node_to_insert def UpperCAmelCase__ ( self :Any , lowercase_ :int , lowercase_ :int ) -> None: UpperCAmelCase = 1 UpperCAmelCase = Node(lowercase_ ) UpperCAmelCase = self.head while node: if current_position == position: self.insert_before_node(lowercase_ , lowercase_ ) return current_position += 1 UpperCAmelCase = node.next self.insert_after_node(self.tail , lowercase_ ) def UpperCAmelCase__ ( self :Optional[Any] , lowercase_ :int ) -> Node: UpperCAmelCase = self.head while node: if node.get_data() == item: return node UpperCAmelCase = node.get_next() raise Exception('Node not found' ) def UpperCAmelCase__ ( self :Any , lowercase_ :Optional[Any] ) -> Dict: if (node := self.get_node(lowercase_ )) is not None: if node == self.head: UpperCAmelCase = self.head.get_next() if node == self.tail: UpperCAmelCase = self.tail.get_previous() self.remove_node_pointers(lowercase_ ) @staticmethod def UpperCAmelCase__ ( lowercase_ :Node ) -> None: if node.get_next(): UpperCAmelCase = node.previous if node.get_previous(): UpperCAmelCase = node.next UpperCAmelCase = None UpperCAmelCase = None def UpperCAmelCase__ ( self :Union[str, Any] ) -> List[str]: return self.head is None def _lowerCAmelCase ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
78
1
'''simple docstring''' lowerCAmelCase : Optional[int] = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} lowerCAmelCase : Union[str, Any] = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : List[Any] = True _lowerCAmelCase : Union[str, Any] = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(_A , _A , _A ) order.append(_A ) return order def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : Tuple = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(_A , _A , _A ) return component def lowercase (_A ): """simple docstring""" _lowerCAmelCase : int = len(_A ) * [False] _lowerCAmelCase : dict[int, list[int]] = {vert: [] for vert in range(len(_A ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(_A ) _lowerCAmelCase : int = [] for i, was_visited in enumerate(_A ): if not was_visited: order += topology_sort(_A , _A , _A ) _lowerCAmelCase : Any = [] _lowerCAmelCase : Union[str, Any] = len(_A ) * [False] for i in range(len(_A ) ): _lowerCAmelCase : Tuple = order[len(_A ) - i - 1] if not visited[vert]: _lowerCAmelCase : Union[str, Any] = find_components(_A , _A , _A ) components_list.append(_A ) return components_list
367
'''simple docstring''' from __future__ import annotations from typing import Any def lowercase (_A ): """simple docstring""" if not postfix_notation: return 0 _lowerCAmelCase : int = {'+', '-', '*', '/'} _lowerCAmelCase : list[Any] = [] for token in postfix_notation: if token in operations: _lowerCAmelCase , _lowerCAmelCase : Tuple = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(_A ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
25
0
__UpperCamelCase : int = frozenset( [ "prompt", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", "cross_attention_kwargs", ] ) __UpperCamelCase : List[Any] = frozenset(["prompt", "negative_prompt"]) __UpperCamelCase : Tuple = frozenset([]) __UpperCamelCase : Union[str, Any] = frozenset(["image"]) __UpperCamelCase : Tuple = frozenset( [ "image", "height", "width", "guidance_scale", ] ) __UpperCamelCase : str = frozenset(["image"]) __UpperCamelCase : Optional[int] = frozenset( [ "prompt", "image", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", ] ) __UpperCamelCase : Union[str, Any] = frozenset(["prompt", "image", "negative_prompt"]) __UpperCamelCase : Dict = frozenset( [ # Text guided image variation with an image mask "prompt", "image", "mask_image", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", ] ) __UpperCamelCase : Any = frozenset(["prompt", "image", "mask_image", "negative_prompt"]) __UpperCamelCase : Dict = frozenset( [ # image variation with an image mask "image", "mask_image", "height", "width", "guidance_scale", ] ) __UpperCamelCase : Any = frozenset(["image", "mask_image"]) __UpperCamelCase : Optional[int] = frozenset( [ "example_image", "image", "mask_image", "height", "width", "guidance_scale", ] ) __UpperCamelCase : Dict = frozenset(["example_image", "image", "mask_image"]) __UpperCamelCase : Union[str, Any] = frozenset(["class_labels"]) __UpperCamelCase : Dict = frozenset(["class_labels"]) __UpperCamelCase : Any = frozenset(["batch_size"]) __UpperCamelCase : Union[str, Any] = frozenset([]) __UpperCamelCase : Any = frozenset(["batch_size"]) __UpperCamelCase : Optional[Any] = frozenset([]) __UpperCamelCase : int = frozenset( [ "prompt", "audio_length_in_s", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", "cross_attention_kwargs", ] ) __UpperCamelCase : Union[str, Any] = frozenset(["prompt", "negative_prompt"]) __UpperCamelCase : str = frozenset(["input_tokens"]) __UpperCamelCase : Any = frozenset(["input_tokens"])
146
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class _UpperCAmelCase ( A__ ): """simple docstring""" def __init__( self : Dict, lowerCamelCase : pyspark.sql.DataFrame, lowerCamelCase : Optional[NamedSplit] = None, lowerCamelCase : Optional[Features] = None, lowerCamelCase : bool = True, lowerCamelCase : str = None, lowerCamelCase : bool = False, lowerCamelCase : str = None, lowerCamelCase : bool = True, lowerCamelCase : str = "arrow", **lowerCamelCase : str, ): '''simple docstring''' super().__init__( split=lowerCamelCase, features=lowerCamelCase, cache_dir=lowerCamelCase, keep_in_memory=lowerCamelCase, streaming=lowerCamelCase, **lowerCamelCase, ) lowercase__ = load_from_cache_file lowercase__ = file_format lowercase__ = Spark( df=lowerCamelCase, features=lowerCamelCase, cache_dir=lowerCamelCase, working_dir=lowerCamelCase, **lowerCamelCase, ) def lowercase__ ( self : List[Any] ): '''simple docstring''' if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowercase__ = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowerCamelCase, file_format=self._file_format, ) return self.builder.as_dataset(split=self.split )
207
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'google/fnet-base': 'https://huggingface.co/google/fnet-base/resolve/main/config.json', 'google/fnet-large': 'https://huggingface.co/google/fnet-large/resolve/main/config.json' # See all FNet models at https://huggingface.co/models?filter=fnet } class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = 'fnet' def __init__( self : Tuple , _UpperCAmelCase : Dict=3_20_00 , _UpperCAmelCase : str=7_68 , _UpperCAmelCase : Union[str, Any]=12 , _UpperCAmelCase : List[str]=30_72 , _UpperCAmelCase : Tuple="gelu_new" , _UpperCAmelCase : int=0.1 , _UpperCAmelCase : List[Any]=5_12 , _UpperCAmelCase : Dict=4 , _UpperCAmelCase : Dict=0.02 , _UpperCAmelCase : List[str]=1E-12 , _UpperCAmelCase : Tuple=False , _UpperCAmelCase : Tuple=5_12 , _UpperCAmelCase : int=3 , _UpperCAmelCase : Optional[Any]=1 , _UpperCAmelCase : Union[str, Any]=2 , **_UpperCAmelCase : Optional[Any] , ): """simple docstring""" super().__init__(pad_token_id=lowercase__ , bos_token_id=lowercase__ , eos_token_id=lowercase__ , **lowercase__ ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = use_tpu_fourier_optimizations UpperCAmelCase__ = tpu_short_seq_length
358
'''simple docstring''' import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ = s.rsplit(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return new.join(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Dict ): '''simple docstring''' return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[str] ): '''simple docstring''' UpperCAmelCase__ = {} UpperCAmelCase__ = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ = key.replace(F'''{group_key}.''' , F'''{group_key}.group.''' ) if "res_path" in key: UpperCAmelCase__ = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): UpperCAmelCase__ = rreplace(SCREAMING_SNAKE_CASE__ , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): UpperCAmelCase__ = rreplace(SCREAMING_SNAKE_CASE__ , """.b""" , """.bias""" , 1 ) UpperCAmelCase__ = value.float() return upgrade @torch.no_grad() def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any=None , SCREAMING_SNAKE_CASE__ : List[Any]=True ): '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ = Encoder() if os.path.exists(SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = torch.load(SCREAMING_SNAKE_CASE__ ) else: UpperCAmelCase__ = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = ckpt.state_dict() encoder.load_state_dict(SCREAMING_SNAKE_CASE__ ) if config_path is not None: UpperCAmelCase__ = FlavaImageCodebookConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) else: UpperCAmelCase__ = FlavaImageCodebookConfig() UpperCAmelCase__ = FlavaImageCodebook(SCREAMING_SNAKE_CASE__ ).eval() UpperCAmelCase__ = encoder.state_dict() UpperCAmelCase__ = upgrade_state_dict(SCREAMING_SNAKE_CASE__ ) hf_model.load_state_dict(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = hf_model.state_dict() UpperCAmelCase__ = count_parameters(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = count_parameters(SCREAMING_SNAKE_CASE__ ) assert torch.allclose(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , atol=1e-3 ) if save_checkpoint: hf_model.save_pretrained(SCREAMING_SNAKE_CASE__ ) else: return hf_state_dict if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to flava checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') UpperCAmelCase_ = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
61
0
import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ : str = inspect.getfile(accelerate.test_utils ) UpperCamelCase__ : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_script.py'''] ) UpperCamelCase__ : str = os.path.sep.join(inspect.getfile(self.__class__ ).split(os.path.sep )[:-1] ) @require_tpu def UpperCAmelCase__ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[Any] = F"\n {self.test_dir}/xla_spawn.py\n --num_cores 8\n {self.test_file_path}\n ".split() UpperCamelCase__ : Tuple = [sys.executable] + distributed_args execute_subprocess_async(_UpperCamelCase , env=os.environ.copy() )
146
'''simple docstring''' import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _snake_case = random.Random() def _A ( snake_case , snake_case=1.0 , snake_case=None , snake_case=None ) -> Optional[Any]: if rng is None: _lowercase : List[str] = global_rng _lowercase : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class a__ ( unittest.TestCase ): def __init__( self , _UpperCamelCase , _UpperCamelCase=7 , _UpperCamelCase=400 , _UpperCamelCase=2000 , _UpperCamelCase=10 , _UpperCamelCase=160 , _UpperCamelCase=8 , _UpperCamelCase=0.0 , _UpperCamelCase=4000 , _UpperCamelCase=False , _UpperCamelCase=True , ): """simple docstring""" _lowercase : int = parent _lowercase : Optional[int] = batch_size _lowercase : List[Any] = min_seq_length _lowercase : Union[str, Any] = max_seq_length _lowercase : int = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowercase : Union[str, Any] = padding_value _lowercase : Dict = sampling_rate _lowercase : Any = return_attention_mask _lowercase : Union[str, Any] = do_normalize _lowercase : int = feature_size _lowercase : str = chunk_length _lowercase : Any = hop_length def _lowerCamelCase ( self ): """simple docstring""" return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowerCamelCase ( self , _UpperCamelCase=False , _UpperCamelCase=False ): """simple docstring""" def _flatten(_UpperCamelCase ): return list(itertools.chain(*_UpperCamelCase ) ) if equal_length: _lowercase : List[Any] = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowercase : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowercase : Optional[Any] = [np.asarray(_UpperCamelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class a__ ( lowerCamelCase_ , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Tuple = WhisperFeatureExtractor if is_speech_available() else None def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Union[str, Any] = WhisperFeatureExtractionTester(self ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowercase : List[Any] = feat_extract_first.save_pretrained(_UpperCamelCase )[0] check_json_file_has_correct_format(_UpperCamelCase ) _lowercase : Tuple = self.feature_extraction_class.from_pretrained(_UpperCamelCase ) _lowercase : List[Any] = feat_extract_first.to_dict() _lowercase : List[str] = feat_extract_second.to_dict() _lowercase : Tuple = feat_extract_first.mel_filters _lowercase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowercase : Optional[int] = os.path.join(_UpperCamelCase , "feat_extract.json" ) feat_extract_first.to_json_file(_UpperCamelCase ) _lowercase : Any = self.feature_extraction_class.from_json_file(_UpperCamelCase ) _lowercase : List[Any] = feat_extract_first.to_dict() _lowercase : str = feat_extract_second.to_dict() _lowercase : List[str] = feat_extract_first.mel_filters _lowercase : Optional[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowercase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] _lowercase : Optional[Any] = [np.asarray(_UpperCamelCase ) for speech_input in speech_inputs] # Test feature size _lowercase : int = feature_extractor(_UpperCamelCase , padding="max_length" , return_tensors="np" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowercase : List[str] = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_features _lowercase : str = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_features self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase , atol=1E-3 ) ) # Test batched _lowercase : Dict = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features _lowercase : Optional[Any] = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_UpperCamelCase , _UpperCamelCase ): self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _lowercase : Optional[int] = [floats_list((1, x) )[0] for x in (800, 800, 800)] _lowercase : List[str] = np.asarray(_UpperCamelCase ) _lowercase : Optional[Any] = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features _lowercase : str = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_UpperCamelCase , _UpperCamelCase ): self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase , atol=1E-3 ) ) # Test truncation required _lowercase : List[Any] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] _lowercase : List[str] = [np.asarray(_UpperCamelCase ) for speech_input in speech_inputs] _lowercase : Any = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowercase : Any = [np.asarray(_UpperCamelCase ) for speech_input in speech_inputs_truncated] _lowercase : List[str] = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features _lowercase : Union[str, Any] = feature_extractor(_UpperCamelCase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_UpperCamelCase , _UpperCamelCase ): self.assertTrue(np.allclose(_UpperCamelCase , _UpperCamelCase , atol=1E-3 ) ) def _lowerCamelCase ( self ): """simple docstring""" import torch _lowercase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowercase : Optional[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) _lowercase : Dict = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowercase : Optional[int] = feature_extractor.pad([{"input_features": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowercase : Optional[int] = feature_extractor.pad([{"input_features": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" _lowercase : int = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech _lowercase : Optional[int] = ds.sort("id" ).select(range(_UpperCamelCase ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def _lowerCamelCase ( self ): """simple docstring""" _lowercase : str = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowercase : str = self._load_datasamples(1 ) _lowercase : Union[str, Any] = WhisperFeatureExtractor() _lowercase : Any = feature_extractor(_UpperCamelCase , return_tensors="pt" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , _UpperCamelCase , atol=1E-4 ) ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowercase : str = self._load_datasamples(1 )[0] _lowercase : List[str] = ((audio - audio.min()) / (audio.max() - audio.min())) * 65535 # Rescale to [0, 65535] to show issue _lowercase : Optional[int] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=_UpperCamelCase )[0] self.assertTrue(np.all(np.mean(_UpperCamelCase ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_UpperCamelCase ) - 1 ) < 1E-3 ) )
250
0
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=30 , __a=2 , __a=3 , __a=True , __a=True , __a=32 , __a=2 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=3 , __a=0.6 , __a=None , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = mask_ratio _UpperCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _UpperCamelCase = (image_size // patch_size) ** 2 _UpperCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1))) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def UpperCAmelCase ( self , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = TFViTMAEModel(config=snake_case__) _UpperCamelCase = model(snake_case__ , training=snake_case__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a) -> Tuple: '''simple docstring''' _UpperCamelCase = TFViTMAEForPreTraining(snake_case__) _UpperCamelCase = model(snake_case__ , training=snake_case__) # expected sequence length = num_patches _UpperCamelCase = (self.image_size // self.patch_size) ** 2 _UpperCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels)) # test greyscale images _UpperCamelCase = 1 _UpperCamelCase = TFViTMAEForPreTraining(snake_case__) _UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) _UpperCamelCase = model(snake_case__ , training=snake_case__) _UpperCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels)) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() (_UpperCamelCase) = config_and_inputs _UpperCamelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): lowercase__ = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = TFViTMAEModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer)) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) _UpperCamelCase = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case__) def UpperCAmelCase ( self) -> int: '''simple docstring''' np.random.seed(2) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = int((config.image_size // config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) _UpperCamelCase = self._prepare_for_class(snake_case__ , snake_case__) _UpperCamelCase = model(snake_case__ , noise=snake_case__) _UpperCamelCase = copy.deepcopy(self._prepare_for_class(snake_case__ , snake_case__)) _UpperCamelCase = model(**snake_case__ , noise=snake_case__) _UpperCamelCase = outputs_dict[0].numpy() _UpperCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords)) , 1e-6) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' np.random.seed(2) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = int((config.image_size // config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) def prepare_numpy_arrays(__a): _UpperCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(snake_case__): _UpperCamelCase = v.numpy() else: _UpperCamelCase = np.array(snake_case__) return inputs_np_dict for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) _UpperCamelCase = self._prepare_for_class(snake_case__ , snake_case__) _UpperCamelCase = prepare_numpy_arrays(snake_case__) _UpperCamelCase = model(snake_case__ , noise=snake_case__) _UpperCamelCase = model(**snake_case__ , noise=snake_case__) self.assert_outputs_same(snake_case__ , snake_case__) def UpperCAmelCase ( self , __a , __a , __a) -> Any: '''simple docstring''' np.random.seed(2) _UpperCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) _UpperCamelCase = tf.constant(snake_case__) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _UpperCamelCase = tf_noise super().check_pt_tf_models(snake_case__ , snake_case__ , snake_case__) def UpperCAmelCase ( self) -> str: '''simple docstring''' np.random.seed(2) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__),) for module_member_name in dir(snake_case__) if module_member_name.endswith('''MainLayer''') # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''')] == model_class.__name__[: -len('''Model''')] for module_member in (getattr(snake_case__ , snake_case__),) if isinstance(snake_case__ , snake_case__) and tf.keras.layers.Layer in module_member.__bases__ and getattr(snake_case__ , '''_keras_serializable''' , snake_case__) } _UpperCamelCase = int((config.image_size // config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) _UpperCamelCase = tf.convert_to_tensor(snake_case__) inputs_dict.update({'''noise''': noise}) for main_layer_class in tf_main_layer_classes: _UpperCamelCase = main_layer_class(snake_case__) _UpperCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype) for name, tensor in inputs_dict.items() } _UpperCamelCase = tf.keras.Model(snake_case__ , outputs=main_layer(snake_case__)) _UpperCamelCase = model(snake_case__) with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = os.path.join(snake_case__ , '''keras_model.h5''') model.save(snake_case__) _UpperCamelCase = tf.keras.models.load_model( snake_case__ , custom_objects={main_layer_class.__name__: main_layer_class}) assert isinstance(snake_case__ , tf.keras.Model) _UpperCamelCase = model(snake_case__) self.assert_outputs_same(snake_case__ , snake_case__) @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' np.random.seed(2) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = int((config.image_size // config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) _UpperCamelCase = self._prepare_for_class(snake_case__ , snake_case__) _UpperCamelCase = model(snake_case__ , noise=snake_case__) if model_class.__name__ == "TFViTMAEModel": _UpperCamelCase = outputs.last_hidden_state.numpy() _UpperCamelCase = 0 else: _UpperCamelCase = outputs.logits.numpy() _UpperCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case__ , saved_model=snake_case__) _UpperCamelCase = model_class.from_pretrained(snake_case__) _UpperCamelCase = model(snake_case__ , noise=snake_case__) if model_class.__name__ == "TFViTMAEModel": _UpperCamelCase = after_outputs['last_hidden_state'].numpy() _UpperCamelCase = 0 else: _UpperCamelCase = after_outputs['logits'].numpy() _UpperCamelCase = 0 _UpperCamelCase = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(snake_case__ , 1e-5) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' np.random.seed(2) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = int((config.image_size // config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) for model_class in self.all_model_classes: _UpperCamelCase = model_class(snake_case__) _UpperCamelCase = self._prepare_for_class(snake_case__ , snake_case__) _UpperCamelCase = model(snake_case__ , noise=snake_case__) _UpperCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(snake_case__) _UpperCamelCase = model_class.from_config(model.get_config()) # make sure it also accepts a normal config _UpperCamelCase = model_class.from_config(model.config) _UpperCamelCase = new_model(snake_case__) # Build model new_model.set_weights(model.get_weights()) _UpperCamelCase = new_model(snake_case__ , noise=snake_case__) self.assert_outputs_same(snake_case__ , snake_case__) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.''') def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''') def UpperCAmelCase ( self) -> Any: '''simple docstring''' pass @slow def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''') self.assertIsNotNone(snake_case__) def lowerCamelCase__ ( ) -> Any: """simple docstring""" _UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''') if is_vision_available() else None @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' np.random.seed(2) _UpperCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''') _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=snake_case__ , return_tensors='''tf''') # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _UpperCamelCase = ViTMAEConfig() _UpperCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2) _UpperCamelCase = np.random.uniform(size=(1, num_patches)) # forward pass _UpperCamelCase = model(**snake_case__ , noise=snake_case__) # verify the logits _UpperCamelCase = tf.convert_to_tensor([1, 1_96, 7_68]) self.assertEqual(outputs.logits.shape , snake_case__) _UpperCamelCase = tf.convert_to_tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]]) tf.debugging.assert_near(outputs.logits[0, :3, :3] , snake_case__ , atol=1e-4)
353
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { """facebook/xlm-roberta-xl""": """https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json""", """facebook/xlm-roberta-xxl""": """https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json""", # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'xlm-roberta-xl' def __init__( self , __a=25_08_80 , __a=25_60 , __a=36 , __a=32 , __a=1_02_40 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_14 , __a=1 , __a=0.02 , __a=1e-05 , __a=1 , __a=0 , __a=2 , __a="absolute" , __a=True , __a=None , **__a , ) -> Tuple: '''simple docstring''' super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = position_embedding_type _UpperCamelCase = use_cache _UpperCamelCase = classifier_dropout class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": _UpperCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ])
100
0
"""simple docstring""" from itertools import permutations def __a ( __lowerCamelCase ): if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False UpperCAmelCase_ : List[str] = [7, 11, 13, 17] for i, test in enumerate(__lowerCamelCase ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def __a ( __lowerCamelCase = 10 ): return sum( int("".join(map(__lowerCamelCase, __lowerCamelCase ) ) ) for num in permutations(range(__lowerCamelCase ) ) if is_substring_divisible(__lowerCamelCase ) ) if __name__ == "__main__": print(f"""{solution() = }""")
61
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig _a = { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/config.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/config.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/config.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/config.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/config.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/config.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json', } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """albert""" def __init__( self , lowercase_=3_0000 , lowercase_=128 , lowercase_=4096 , lowercase_=12 , lowercase_=1 , lowercase_=64 , lowercase_=1_6384 , lowercase_=1 , lowercase_="gelu_new" , lowercase_=0 , lowercase_=0 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1E-1_2 , lowercase_=0.1 , lowercase_="absolute" , lowercase_=0 , lowercase_=2 , lowercase_=3 , **lowercase_ , ): """simple docstring""" super().__init__(pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , **lowercase_ ) UpperCAmelCase_ : int = vocab_size UpperCAmelCase_ : Optional[int] = embedding_size UpperCAmelCase_ : List[str] = hidden_size UpperCAmelCase_ : Optional[int] = num_hidden_layers UpperCAmelCase_ : Union[str, Any] = num_hidden_groups UpperCAmelCase_ : Dict = num_attention_heads UpperCAmelCase_ : Any = inner_group_num UpperCAmelCase_ : Union[str, Any] = hidden_act UpperCAmelCase_ : Union[str, Any] = intermediate_size UpperCAmelCase_ : List[str] = hidden_dropout_prob UpperCAmelCase_ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase_ : Optional[Any] = max_position_embeddings UpperCAmelCase_ : Any = type_vocab_size UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : List[Any] = classifier_dropout_prob UpperCAmelCase_ : Tuple = position_embedding_type class A_ (lowercase__ ): '''simple docstring''' @property def UpperCamelCase__ ( self ): """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase_ : int = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
61
1
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=32 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=[10, 20, 30, 40] , SCREAMING_SNAKE_CASE__=[1, 1, 2, 1] , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__="relu" , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=None , ): lowercase : Any = parent lowercase : Optional[Any] = batch_size lowercase : str = image_size lowercase : Any = num_channels lowercase : Any = embeddings_size lowercase : Optional[int] = hidden_sizes lowercase : Dict = depths lowercase : Optional[int] = is_training lowercase : List[str] = use_labels lowercase : List[Any] = hidden_act lowercase : Union[str, Any] = num_labels lowercase : str = scope lowercase : Optional[int] = len(SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self ): lowercase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase : Any = None if self.use_labels: lowercase : Dict = ids_tensor([self.batch_size] , self.num_labels ) lowercase : List[Any] = self.get_config() return config, pixel_values, labels def __lowerCamelCase ( self ): return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase : Union[str, Any] = RegNetModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() lowercase : str = model(SCREAMING_SNAKE_CASE__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase : Dict = self.num_labels lowercase : Optional[int] = RegNetForImageClassification(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() lowercase : Dict = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCamelCase ( self ): lowercase : List[Any] = self.prepare_config_and_inputs() lowercase , lowercase , lowercase : List[str] = config_and_inputs lowercase : str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( A__ , A__ , unittest.TestCase ): A : str = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () A : Union[str, Any] = ( {'feature-extraction': RegNetModel, 'image-classification': RegNetForImageClassification} if is_torch_available() else {} ) A : List[str] = False A : str = False A : int = False A : Dict = False def __lowerCamelCase ( self ): lowercase : Dict = RegNetModelTester(self ) lowercase : str = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCamelCase ( self ): return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def __lowerCamelCase ( self ): pass def __lowerCamelCase ( self ): lowercase , lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : str = model_class(SCREAMING_SNAKE_CASE__ ) lowercase : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : Optional[Any] = [*signature.parameters.keys()] lowercase : Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self ): lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self ): lowercase , lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Any = model_class(config=SCREAMING_SNAKE_CASE__ ) for name, module in model.named_modules(): if isinstance(SCREAMING_SNAKE_CASE__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def __lowerCamelCase ( self ): def check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase : Tuple = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): lowercase : int = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowercase : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase : Dict = self.model_tester.num_stages self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) lowercase , lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase : List[str] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase : Optional[int] = layer_type lowercase : List[str] = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase : Optional[Any] = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self ): lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE__ ) @slow def __lowerCamelCase ( self ): for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : Any = RegNetModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def __lowercase ( ) ->Tuple: """simple docstring""" lowercase : Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def __lowerCamelCase ( self ): return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __lowerCamelCase ( self ): lowercase : Optional[Any] = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(SCREAMING_SNAKE_CASE__ ) lowercase : Optional[int] = self.default_image_processor lowercase : List[Any] = prepare_img() lowercase : Any = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): lowercase : Tuple = model(**SCREAMING_SNAKE_CASE__ ) # verify the logits lowercase : Optional[int] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE__ ) lowercase : List[Any] = torch.tensor([-0.4180, -1.5051, -3.4836] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) )
173
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __a = 16 __a = 32 def __lowercase ( _UpperCamelCase, _UpperCamelCase = 16 ) ->List[Any]: """simple docstring""" lowercase : Optional[Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowercase : List[Any] = load_dataset('''glue''', '''mrpc''' ) def tokenize_function(_UpperCamelCase ): # max_length=None => use the model max length (it's actually the default) lowercase : List[Any] = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=_UpperCamelCase, max_length=_UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase : Union[str, Any] = datasets.map( _UpperCamelCase, batched=_UpperCamelCase, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase : Union[str, Any] = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(_UpperCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase : Optional[Any] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase : Tuple = 16 elif accelerator.mixed_precision != "no": lowercase : str = 8 else: lowercase : List[str] = None return tokenizer.pad( _UpperCamelCase, padding='''longest''', max_length=_UpperCamelCase, pad_to_multiple_of=_UpperCamelCase, return_tensors='''pt''', ) # Instantiate dataloaders. lowercase : int = DataLoader( tokenized_datasets['''train'''], shuffle=_UpperCamelCase, collate_fn=_UpperCamelCase, batch_size=_UpperCamelCase ) lowercase : str = DataLoader( tokenized_datasets['''validation'''], shuffle=_UpperCamelCase, collate_fn=_UpperCamelCase, batch_size=_UpperCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders __a = mocked_dataloaders # noqa: F811 def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->str: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''', _UpperCamelCase ) == "1": lowercase : Tuple = 2 # New Code # lowercase : Optional[int] = int(args.gradient_accumulation_steps ) lowercase : Optional[int] = int(args.local_sgd_steps ) # Initialize accelerator lowercase : Tuple = Accelerator( cpu=args.cpu, mixed_precision=args.mixed_precision, gradient_accumulation_steps=_UpperCamelCase ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('''LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase : Dict = config['''lr'''] lowercase : List[str] = int(config['''num_epochs'''] ) lowercase : str = int(config['''seed'''] ) lowercase : str = int(config['''batch_size'''] ) lowercase : Any = evaluate.load('''glue''', '''mrpc''' ) set_seed(_UpperCamelCase ) lowercase , lowercase : Dict = get_dataloaders(_UpperCamelCase, _UpperCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase : Optional[int] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''', return_dict=_UpperCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase : int = model.to(accelerator.device ) # Instantiate optimizer lowercase : Any = AdamW(params=model.parameters(), lr=_UpperCamelCase ) # Instantiate scheduler lowercase : Union[str, Any] = get_linear_schedule_with_warmup( optimizer=_UpperCamelCase, num_warmup_steps=100, num_training_steps=(len(_UpperCamelCase ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase , lowercase , lowercase , lowercase , lowercase : Optional[Any] = accelerator.prepare( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) # Now we train the model for epoch in range(_UpperCamelCase ): model.train() with LocalSGD( accelerator=_UpperCamelCase, model=_UpperCamelCase, local_sgd_steps=_UpperCamelCase, enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(_UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_UpperCamelCase ): lowercase : int = model(**_UpperCamelCase ) lowercase : Optional[int] = output.loss accelerator.backward(_UpperCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(_UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase : Optional[int] = model(**_UpperCamelCase ) lowercase : Optional[Any] = outputs.logits.argmax(dim=-1 ) lowercase , lowercase : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=_UpperCamelCase, references=_UpperCamelCase, ) lowercase : int = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""", _UpperCamelCase ) def __lowercase ( ) ->int: """simple docstring""" lowercase : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''', type=_UpperCamelCase, default=_UpperCamelCase, choices=['''no''', '''fp16''', '''bf16''', '''fp8'''], help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''', ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''', type=_UpperCamelCase, default=1, help='''The number of minibatches to be ran before gradients are accumulated.''', ) parser.add_argument( '''--local_sgd_steps''', type=_UpperCamelCase, default=8, help='''Number of local SGD steps or None to disable local SGD''' ) parser.add_argument('''--cpu''', action='''store_true''', help='''If passed, will train on the CPU.''' ) lowercase : List[Any] = parser.parse_args() lowercase : List[Any] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(_UpperCamelCase, _UpperCamelCase ) if __name__ == "__main__": main()
173
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=False ) -> Any: '''simple docstring''' lowercase_ = [] # fmt: off # stem: rename_keys.append(("""cls_token""", """vit.embeddings.cls_token""") ) rename_keys.append(("""pos_embed""", """vit.embeddings.position_embeddings""") ) rename_keys.append(("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias""") ) # backbone rename_keys.append(("""patch_embed.backbone.stem.conv.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.bias""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias""") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowercase_ = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) # fmt: on return rename_keys def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: lowercase_ = """""" else: lowercase_ = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase_ = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) lowercase_ = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase_ = in_proj_weight[ : config.hidden_size, : ] lowercase_ = in_proj_bias[: config.hidden_size] lowercase_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase_ = in_proj_weight[ -config.hidden_size :, : ] lowercase_ = in_proj_bias[-config.hidden_size :] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(__snake_case , __snake_case ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' lowercase_ = dct.pop(__snake_case ) lowercase_ = val def _SCREAMING_SNAKE_CASE () -> Tuple: '''simple docstring''' lowercase_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowercase_ = Image.open(requests.get(__snake_case , stream=__snake_case ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> str: '''simple docstring''' lowercase_ = BitConfig( global_padding="""same""" , layer_type="""bottleneck""" , depths=(3, 4, 9) , out_features=["""stage3"""] , embedding_dynamic_padding=__snake_case , ) lowercase_ = ViTHybridConfig(backbone_config=__snake_case , image_size=3_84 , num_labels=10_00 ) lowercase_ = False # load original model from timm lowercase_ = timm.create_model(__snake_case , pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowercase_ = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) lowercase_ = create_rename_keys(__snake_case , __snake_case ) for src, dest in rename_keys: rename_key(__snake_case , __snake_case , __snake_case ) read_in_q_k_v(__snake_case , __snake_case , __snake_case ) lowercase_ = """huggingface/label-files""" lowercase_ = """imagenet-1k-id2label.json""" lowercase_ = json.load(open(hf_hub_download(__snake_case , __snake_case , repo_type="""dataset""" ) , """r""" ) ) lowercase_ = {int(__snake_case ): v for k, v in idalabel.items()} lowercase_ = idalabel lowercase_ = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": lowercase_ = ViTHybridModel(__snake_case ).eval() else: lowercase_ = ViTHybridForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # create image processor lowercase_ = create_transform(**resolve_data_config({} , model=__snake_case ) ) lowercase_ = transform.transforms lowercase_ = { """bilinear""": PILImageResampling.BILINEAR, """bicubic""": PILImageResampling.BICUBIC, """nearest""": PILImageResampling.NEAREST, } lowercase_ = ViTHybridImageProcessor( do_resize=__snake_case , size={"""shortest_edge""": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=__snake_case , crop_size={"""height""": timm_transforms[1].size[0], """width""": timm_transforms[1].size[1]} , do_normalize=__snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase_ = prepare_img() lowercase_ = transform(__snake_case ).unsqueeze(0 ) lowercase_ = processor(__snake_case , return_tensors="""pt""" ).pixel_values # verify pixel values assert torch.allclose(__snake_case , __snake_case ) # verify logits with torch.no_grad(): lowercase_ = model(__snake_case ) lowercase_ = outputs.logits print("""Predicted class:""" , logits.argmax(-1 ).item() ) if base_model: lowercase_ = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case , outputs.pooler_output , atol=1E-3 ) else: lowercase_ = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case , outputs.logits , atol=1E-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(__snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) UpperCAmelCase : Optional[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
136
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = { "configuration_lxmert": ["LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LxmertConfig"], "tokenization_lxmert": ["LxmertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["LxmertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ "LxmertEncoder", "LxmertForPreTraining", "LxmertForQuestionAnswering", "LxmertModel", "LxmertPreTrainedModel", "LxmertVisualFeatureEncoder", "LxmertXLayer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ "TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLxmertForPreTraining", "TFLxmertMainLayer", "TFLxmertModel", "TFLxmertPreTrainedModel", "TFLxmertVisualFeatureEncoder", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys _a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
209
0
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 UpperCamelCase = get_tests_dir('fixtures') UpperCamelCase = get_tests_dir('fixtures/dummy_feature_extractor_config.json') UpperCamelCase = get_tests_dir('fixtures/dummy-config.json') class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" def a ( self : Union[str, Any] ) -> str: lowerCAmelCase__ = 0 def a ( self : Optional[int] ) -> Optional[Any]: lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : Tuple ) -> Dict: lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] ) -> Tuple: with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ).to_dict() config_dict.pop("feature_extractor_type" ) lowerCAmelCase__ = WavaVecaFeatureExtractor(**SCREAMING_SNAKE_CASE__ ) # save in new folder model_config.save_pretrained(SCREAMING_SNAKE_CASE__ ) config.save_pretrained(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ) # make sure private variable is not incorrectly saved lowerCAmelCase__ = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : List[str] ) -> Dict: lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] ) -> Dict: with self.assertRaisesRegex( SCREAMING_SNAKE_CASE__ , "bert-base is not a local folder and is not a valid model identifier" ): lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained("bert-base" ) def a ( self : Tuple ) -> str: with self.assertRaisesRegex( SCREAMING_SNAKE_CASE__ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ , revision="aaaaaa" ) def a ( self : int ) -> Dict: with self.assertRaisesRegex( SCREAMING_SNAKE_CASE__ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def a ( self : List[Any] ) -> List[Any]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=SCREAMING_SNAKE_CASE__ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ , trust_remote_code=SCREAMING_SNAKE_CASE__ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def a ( self : int ) -> Tuple: try: AutoConfig.register("custom" , SCREAMING_SNAKE_CASE__ ) AutoFeatureExtractor.register(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(SCREAMING_SNAKE_CASE__ ): AutoFeatureExtractor.register(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCAmelCase__ = CustomFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def a ( self : Tuple ) -> Dict: class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = True try: AutoConfig.register("custom" , SCREAMING_SNAKE_CASE__ ) AutoFeatureExtractor.register(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # If remote code is not set, the default is to use local lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=SCREAMING_SNAKE_CASE__ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub lowerCAmelCase__ = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=SCREAMING_SNAKE_CASE__ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(SCREAMING_SNAKE_CASE__ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
366
from statistics import mean, stdev def _A ( lowerCAmelCase_ : list , lowerCAmelCase_ : int = 3 ): """simple docstring""" lowerCAmelCase__ = min(lowerCAmelCase_ ) lowerCAmelCase__ = max(lowerCAmelCase_ ) # normalize data return [round((x - x_min) / (x_max - x_min) , lowerCAmelCase_ ) for x in data] def _A ( lowerCAmelCase_ : list , lowerCAmelCase_ : int = 3 ): """simple docstring""" lowerCAmelCase__ = mean(lowerCAmelCase_ ) lowerCAmelCase__ = stdev(lowerCAmelCase_ ) # standardize data return [round((x - mu) / (sigma) , lowerCAmelCase_ ) for x in data]
221
0
import torch from transformers import AutoModel class __SCREAMING_SNAKE_CASE ( torch.nn.Module ): def __init__( self : Tuple , A : Optional[int]="sayef/fsner-bert-base-uncased" ) ->Any: super(A , self ).__init__() lowerCamelCase__ : Union[str, Any] = AutoModel.from_pretrained(A , return_dict=A ) lowerCamelCase__ : Optional[int] = torch.nn.CosineSimilarity(3 , 1e-08 ) lowerCamelCase__ : str = torch.nn.Softmax(dim=1 ) def __lowerCamelCase ( self : int , **A : Optional[int] ) ->Optional[int]: return self.bert(**A ).last_hidden_state def __lowerCamelCase ( self : Optional[int] , A : List[str] ) ->str: return token_embeddings.sum(2 , keepdim=A ) def __lowerCamelCase ( self : Optional[int] , A : List[Any] , A : List[Any] , A : Optional[int]=1 ) ->Union[str, Any]: return self.softmax(T * self.cos(A , A ) ) def __lowerCamelCase ( self : Union[str, Any] , A : List[str] , A : Optional[int] ) ->str: lowerCamelCase__ : int = W_supports['''sizes'''].tolist() lowerCamelCase__ : List[Any] = W_supports['''start_token_id'''].item() lowerCamelCase__ : Optional[Any] = W_supports['''end_token_id'''].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowerCamelCase__ : str = self.BERT(**A ) lowerCamelCase__ : Optional[int] = self.BERT(**A ) lowerCamelCase__ : List[str] = None lowerCamelCase__ : Dict = None lowerCamelCase__ : Dict = W_supports['''input_ids'''] == start_token_id lowerCamelCase__ : Optional[Any] = W_supports['''input_ids'''] == end_token_id for i, size in enumerate(A ): if i == 0: lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : Optional[Any] = support_sizes[i - 1] lowerCamelCase__ : Dict = S[s : s + size][start_token_masks[s : s + size]] lowerCamelCase__ : Tuple = S[s : s + size][end_token_masks[s : s + size]] lowerCamelCase__ : Optional[int] = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) lowerCamelCase__ : Union[str, Any] = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowerCamelCase__ : Tuple = torch.vstack((p_starts, p_start) ) lowerCamelCase__ : Optional[Any] = torch.vstack((p_ends, p_end) ) else: lowerCamelCase__ : Any = p_start lowerCamelCase__ : List[str] = p_end return p_starts, p_ends
142
from collections import namedtuple import requests from lxml import html # type: ignore _A : Any = namedtuple('covid_data', 'cases deaths recovered') def _a ( UpperCAmelCase = "https://www.worldometers.info/coronavirus/" ) -> covid_data: """simple docstring""" lowerCamelCase__ : Optional[Any] = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(UpperCAmelCase ).content ).xpath(UpperCAmelCase ) ) _A : Dict = 'Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}' print(fmt.format(*covid_stats()))
142
1
from __future__ import annotations class lowerCAmelCase__ : def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : Tuple=None ) -> Optional[int]: __lowerCamelCase = data __lowerCamelCase = None def __repr__( self : str ) -> List[str]: __lowerCamelCase = [] __lowerCamelCase = self while temp: string_rep.append(f'''{temp.data}''' ) __lowerCamelCase = temp.next return "->".join(lowerCamelCase__ ) def __magic_name__ ( __lowerCAmelCase : list ) -> Optional[int]: """simple docstring""" if not elements_list: raise Exception('''The Elements List is empty''' ) __lowerCamelCase = Node(elements_list[0] ) for i in range(1 , len(__lowerCAmelCase ) ): __lowerCamelCase = Node(elements_list[i] ) __lowerCamelCase = current.next return head def __magic_name__ ( __lowerCAmelCase : Node ) -> str: """simple docstring""" if head_node is not None and isinstance(__lowerCAmelCase , __lowerCAmelCase ): print_reverse(head_node.next ) print(head_node.data ) def __magic_name__ ( ) -> Union[str, Any]: """simple docstring""" from doctest import testmod testmod() __lowerCamelCase = make_linked_list([14, 52, 14, 12, 43] ) print('''Linked List:''' ) print(__lowerCAmelCase ) print('''Elements in Reverse:''' ) print_reverse(__lowerCAmelCase ) if __name__ == "__main__": main()
357
from functools import lru_cache def __magic_name__ ( __lowerCAmelCase : int ) -> set: __lowerCamelCase = 2 __lowerCamelCase = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(__lowerCAmelCase ) if n > 1: factors.add(__lowerCAmelCase ) return factors @lru_cache def __magic_name__ ( __lowerCAmelCase : int ) -> int: return len(unique_prime_factors(__lowerCAmelCase ) ) def __magic_name__ ( __lowerCAmelCase : list ) -> bool: return len(set(__lowerCAmelCase ) ) in (0, 1) def __magic_name__ ( __lowerCAmelCase : int ) -> list: __lowerCamelCase = 2 while True: # Increment each value of a generated range __lowerCamelCase = [base + i for i in range(__lowerCAmelCase )] # Run elements through out unique_prime_factors function # Append our target number to the end. __lowerCamelCase = [upf_len(__lowerCAmelCase ) for x in group] checker.append(__lowerCAmelCase ) # If all numbers in the list are equal, return the group variable. if equality(__lowerCAmelCase ): return group # Increment our base variable by 1 base += 1 def __magic_name__ ( __lowerCAmelCase : int = 4 ) -> int: __lowerCamelCase = run(__lowerCAmelCase ) return results[0] if len(__lowerCAmelCase ) else None if __name__ == "__main__": print(solution())
339
0
'''simple docstring''' def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): __UpperCamelCase : Dict = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: __UpperCamelCase : str = 1 - (matter_density + radiation_density + dark_energy) __UpperCamelCase : List[Any] = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) __UpperCamelCase : Optional[Any] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _lowerCAmelCase = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
298
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig _lowerCAmelCase = logging.get_logger(__name__) # General docstring _lowerCAmelCase = '''RegNetConfig''' # Base docstring _lowerCAmelCase = '''facebook/regnet-y-040''' _lowerCAmelCase = [1, 1088, 7, 7] # Image classification docstring _lowerCAmelCase = '''facebook/regnet-y-040''' _lowerCAmelCase = '''tabby, tabby cat''' _lowerCAmelCase = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase = 3 , _UpperCAmelCase = 1 , _UpperCAmelCase = 1 , _UpperCAmelCase = "relu" , **_UpperCAmelCase , ) -> Optional[int]: super().__init__(**_UpperCAmelCase ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb __UpperCamelCase : List[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) __UpperCamelCase : Tuple = tf.keras.layers.ConvaD( filters=_UpperCAmelCase , kernel_size=_UpperCAmelCase , strides=_UpperCAmelCase , padding="VALID" , groups=_UpperCAmelCase , use_bias=_UpperCAmelCase , name="convolution" , ) __UpperCamelCase : int = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) __UpperCamelCase : List[str] = ACTaFN[activation] if activation is not None else tf.identity def a_ (self , _UpperCAmelCase ) -> Dict: __UpperCamelCase : str = self.convolution(self.padding(_UpperCAmelCase ) ) __UpperCamelCase : Dict = self.normalization(_UpperCAmelCase ) __UpperCamelCase : Dict = self.activation(_UpperCAmelCase ) return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , **_UpperCAmelCase ) -> Optional[Any]: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : Any = config.num_channels __UpperCamelCase : str = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def a_ (self , _UpperCAmelCase ) -> Tuple: __UpperCamelCase : Dict = shape_list(_UpperCAmelCase )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) __UpperCamelCase : Any = tf.transpose(_UpperCAmelCase , perm=(0, 2, 3, 1) ) __UpperCamelCase : List[Any] = self.embedder(_UpperCAmelCase ) return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase = 2 , **_UpperCAmelCase ) -> Any: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : Any = tf.keras.layers.ConvaD( filters=_UpperCAmelCase , kernel_size=1 , strides=_UpperCAmelCase , use_bias=_UpperCAmelCase , name="convolution" ) __UpperCamelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase = False ) -> tf.Tensor: return self.normalization(self.convolution(_UpperCAmelCase ) , training=_UpperCAmelCase ) class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ) -> Any: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=_UpperCAmelCase , name="pooler" ) __UpperCamelCase : Optional[Any] = [ tf.keras.layers.ConvaD(filters=_UpperCAmelCase , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=_UpperCAmelCase , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def a_ (self , _UpperCAmelCase ) -> Tuple: # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] __UpperCamelCase : List[str] = self.pooler(_UpperCAmelCase ) for layer_module in self.attention: __UpperCamelCase : str = layer_module(_UpperCAmelCase ) __UpperCamelCase : List[Any] = hidden_state * pooled return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 , **_UpperCAmelCase ) -> int: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : List[Any] = in_channels != out_channels or stride != 1 __UpperCamelCase : List[str] = max(1 , out_channels // config.groups_width ) __UpperCamelCase : List[Any] = ( TFRegNetShortCut(_UpperCAmelCase , stride=_UpperCAmelCase , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. __UpperCamelCase : Optional[Any] = [ TFRegNetConvLayer(_UpperCAmelCase , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(_UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase , name="layer.2" ), ] __UpperCamelCase : Dict = ACTaFN[config.hidden_act] def a_ (self , _UpperCAmelCase ) -> Union[str, Any]: __UpperCamelCase : List[Any] = hidden_state for layer_module in self.layers: __UpperCamelCase : Dict = layer_module(_UpperCAmelCase ) __UpperCamelCase : List[Any] = self.shortcut(_UpperCAmelCase ) hidden_state += residual __UpperCamelCase : Tuple = self.activation(_UpperCAmelCase ) return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 , **_UpperCAmelCase ) -> Any: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : str = in_channels != out_channels or stride != 1 __UpperCamelCase : Optional[int] = max(1 , out_channels // config.groups_width ) __UpperCamelCase : Union[str, Any] = ( TFRegNetShortCut(_UpperCAmelCase , stride=_UpperCAmelCase , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) __UpperCamelCase : Union[str, Any] = [ TFRegNetConvLayer(_UpperCAmelCase , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(_UpperCAmelCase , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(_UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase , name="layer.3" ), ] __UpperCamelCase : Union[str, Any] = ACTaFN[config.hidden_act] def a_ (self , _UpperCAmelCase ) -> int: __UpperCamelCase : str = hidden_state for layer_module in self.layers: __UpperCamelCase : Any = layer_module(_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = self.shortcut(_UpperCAmelCase ) hidden_state += residual __UpperCamelCase : Union[str, Any] = self.activation(_UpperCAmelCase ) return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 2 , _UpperCAmelCase = 2 , **_UpperCAmelCase ) -> int: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : List[str] = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer __UpperCamelCase : Tuple = [ # downsampling is done in the first layer with stride of 2 layer(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , name="layers.0" ), *[layer(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , name=f"layers.{i+1}" ) for i in range(depth - 1 )], ] def a_ (self , _UpperCAmelCase ) -> Any: for layer_module in self.layers: __UpperCamelCase : Dict = layer_module(_UpperCAmelCase ) return hidden_state class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self , _UpperCAmelCase , **_UpperCAmelCase ) -> str: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : Dict = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( _UpperCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) __UpperCamelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(_UpperCAmelCase , config.depths[1:] ) ): self.stages.append(TFRegNetStage(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , depth=_UpperCAmelCase , name=f"stages.{i+1}" ) ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase = False , _UpperCAmelCase = True ) -> TFBaseModelOutputWithNoAttention: __UpperCamelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __UpperCamelCase : Any = hidden_states + (hidden_state,) __UpperCamelCase : Any = stage_module(_UpperCAmelCase ) if output_hidden_states: __UpperCamelCase : List[Any] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=_UpperCAmelCase , hidden_states=_UpperCAmelCase ) @keras_serializable class A ( tf.keras.layers.Layer ): '''simple docstring''' A = RegNetConfig def __init__(self , _UpperCAmelCase , **_UpperCAmelCase ) -> List[Any]: super().__init__(**_UpperCAmelCase ) __UpperCamelCase : Optional[int] = config __UpperCamelCase : List[Any] = TFRegNetEmbeddings(_UpperCAmelCase , name="embedder" ) __UpperCamelCase : Union[str, Any] = TFRegNetEncoder(_UpperCAmelCase , name="encoder" ) __UpperCamelCase : Optional[Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=_UpperCAmelCase , name="pooler" ) @unpack_inputs def a_ (self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: __UpperCamelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase : Dict = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase : Union[str, Any] = self.embedder(_UpperCAmelCase , training=_UpperCAmelCase ) __UpperCamelCase : str = self.encoder( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase , training=_UpperCAmelCase ) __UpperCamelCase : List[str] = encoder_outputs[0] __UpperCamelCase : Tuple = self.pooler(_UpperCAmelCase ) # Change to NCHW output format have uniformity in the modules __UpperCamelCase : List[str] = tf.transpose(_UpperCAmelCase , perm=(0, 3, 1, 2) ) __UpperCamelCase : List[Any] = tf.transpose(_UpperCAmelCase , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: __UpperCamelCase : List[str] = tuple([tf.transpose(_UpperCAmelCase , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_UpperCAmelCase , pooler_output=_UpperCAmelCase , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A = RegNetConfig A = "regnet" A = "pixel_values" @property def a_ (self ) -> List[Any]: return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_2_4, 2_2_4) , dtype=tf.floataa )} _lowerCAmelCase = R''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' _lowerCAmelCase = R''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) -> Tuple: super().__init__(_UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = TFRegNetMainLayer(_UpperCAmelCase , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: __UpperCamelCase : List[str] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase : Optional[int] = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase : Tuple = self.regnet( pixel_values=_UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase , training=_UpperCAmelCase , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) -> int: super().__init__(_UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = config.num_labels __UpperCamelCase : Any = TFRegNetMainLayer(_UpperCAmelCase , name="regnet" ) # classification head __UpperCamelCase : List[str] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a_ (self , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: __UpperCamelCase : Dict = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase : str = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase : Dict = self.regnet( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase , training=_UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] __UpperCamelCase : List[str] = self.classifier[0](_UpperCAmelCase ) __UpperCamelCase : Optional[int] = self.classifier[1](_UpperCAmelCase ) __UpperCamelCase : str = None if labels is None else self.hf_compute_loss(labels=_UpperCAmelCase , logits=_UpperCAmelCase ) if not return_dict: __UpperCamelCase : Union[str, Any] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=_UpperCAmelCase , logits=_UpperCAmelCase , hidden_states=outputs.hidden_states )
298
1
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex A : List[Any] = logging.getLogger(__name__) class A : '''simple docstring''' def __init__( self : Any ) -> int: """simple docstring""" A__ = False def a_ ( self : List[str] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Dict , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[int] ) -> List[str]: """simple docstring""" if not self.initialized: A__ = RagRetriever( __lowerCAmelCase , question_encoder_tokenizer=__lowerCAmelCase , generator_tokenizer=__lowerCAmelCase , index=__lowerCAmelCase , init_retrieval=__lowerCAmelCase , ) A__ = True def a_ ( self : str ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def a_ ( self : Union[str, Any] , __lowerCAmelCase : int , __lowerCAmelCase : Tuple ) -> Tuple: """simple docstring""" A__ , A__ = self.retriever._main_retrieve(__lowerCAmelCase , __lowerCAmelCase ) return doc_ids, retrieved_doc_embeds class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : int , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any]=None ) -> Tuple: """simple docstring""" if index is not None and index.is_initialized() and len(__lowerCAmelCase ) > 0: raise ValueError( """When using Ray for distributed fine-tuning, """ """you'll need to provide the paths instead, """ """as the dataset and the index are loaded """ """separately. More info in examples/rag/use_own_knowledge_dataset.py """ ) super().__init__( __lowerCAmelCase , question_encoder_tokenizer=__lowerCAmelCase , generator_tokenizer=__lowerCAmelCase , index=__lowerCAmelCase , init_retrieval=__lowerCAmelCase , ) A__ = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) for worker in self.retrieval_workers ] ) def a_ ( self : Tuple ) -> Optional[Any]: """simple docstring""" logger.info("""initializing retrieval""" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def a_ ( self : Tuple , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str] ) -> Tuple: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. A__ = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] A__ , A__ = ray.get(random_worker.retrieve.remote(__lowerCAmelCase , __lowerCAmelCase ) ) else: A__ , A__ = self._main_retrieve(__lowerCAmelCase , __lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(__lowerCAmelCase ) @classmethod def a_ ( cls : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Any=None , **__lowerCAmelCase : Any ) -> Tuple: """simple docstring""" return super(__lowerCAmelCase , cls ).get_tokenizers(__lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ) @classmethod def a_ ( cls : Tuple , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : Optional[Any]=None , **__lowerCAmelCase : List[Any] ) -> List[str]: """simple docstring""" A__ = kwargs.pop("""config""" , __lowerCAmelCase ) or RagConfig.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase ) A__ = RagTokenizer.from_pretrained(__lowerCAmelCase , config=__lowerCAmelCase ) A__ = rag_tokenizer.question_encoder A__ = rag_tokenizer.generator if indexed_dataset is not None: A__ = """custom""" A__ = CustomHFIndex(config.retrieval_vector_size , __lowerCAmelCase ) else: A__ = cls._build_index(__lowerCAmelCase ) return cls( __lowerCAmelCase , question_encoder_tokenizer=__lowerCAmelCase , generator_tokenizer=__lowerCAmelCase , retrieval_workers=__lowerCAmelCase , index=__lowerCAmelCase , )
276
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Tuple = logging.get_logger(__name__) A : Optional[int] = { '''roberta-base''': '''https://huggingface.co/roberta-base/resolve/main/config.json''', '''roberta-large''': '''https://huggingface.co/roberta-large/resolve/main/config.json''', '''roberta-large-mnli''': '''https://huggingface.co/roberta-large-mnli/resolve/main/config.json''', '''distilroberta-base''': '''https://huggingface.co/distilroberta-base/resolve/main/config.json''', '''roberta-base-openai-detector''': '''https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json''', '''roberta-large-openai-detector''': '''https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json''', } class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Union[str, Any] = '''roberta''' def __init__( self : Any , __lowerCAmelCase : Tuple=5_02_65 , __lowerCAmelCase : Optional[int]=7_68 , __lowerCAmelCase : Union[str, Any]=12 , __lowerCAmelCase : Dict=12 , __lowerCAmelCase : Optional[Any]=30_72 , __lowerCAmelCase : Any="gelu" , __lowerCAmelCase : Optional[Any]=0.1 , __lowerCAmelCase : Optional[int]=0.1 , __lowerCAmelCase : Optional[Any]=5_12 , __lowerCAmelCase : int=2 , __lowerCAmelCase : List[str]=0.0_2 , __lowerCAmelCase : Dict=1e-12 , __lowerCAmelCase : Optional[int]=1 , __lowerCAmelCase : Tuple=0 , __lowerCAmelCase : int=2 , __lowerCAmelCase : Dict="absolute" , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : Union[str, Any]=None , **__lowerCAmelCase : str , ) -> str: """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_act A__ = intermediate_size A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = initializer_range A__ = layer_norm_eps A__ = position_embedding_type A__ = use_cache A__ = classifier_dropout class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def a_ ( self : int ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": A__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
276
1
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __lowercase : """simple docstring""" _UpperCAmelCase : List[str] _UpperCAmelCase : Optional[str] = None # Automatically constructed _UpperCAmelCase : ClassVar[str] = "dict" _UpperCAmelCase : ClassVar[Any] = None _UpperCAmelCase : str = field(default='''Translation''' , init=UpperCAmelCase_ , repr=UpperCAmelCase_ ) def __call__( self : str): return pa.struct({lang: pa.string() for lang in sorted(self.languages)}) def _SCREAMING_SNAKE_CASE ( self : Tuple): from .features import Value return {k: Value("string") for k in sorted(self.languages)} @dataclass class __lowercase : """simple docstring""" _UpperCAmelCase : Optional[List] = None _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Optional[str] = None # Automatically constructed _UpperCAmelCase : ClassVar[str] = "dict" _UpperCAmelCase : ClassVar[Any] = None _UpperCAmelCase : str = field(default='''TranslationVariableLanguages''' , init=UpperCAmelCase_ , repr=UpperCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: List[str] = sorted(set(self.languages)) if self.languages else None SCREAMING_SNAKE_CASE_: Tuple = len(self.languages) if self.languages else None def __call__( self : Any): return pa.struct({"language": pa.list_(pa.string()), "translation": pa.list_(pa.string())}) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : Any): SCREAMING_SNAKE_CASE_: Optional[Any] = set(self.languages) if self.languages and set(lowerCAmelCase__) - lang_set: raise ValueError( F"Some languages in example ({', '.join(sorted(set(lowerCAmelCase__) - lang_set))}) are not in valid set ({', '.join(lowerCAmelCase__)}).") # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. SCREAMING_SNAKE_CASE_: Tuple = [] for lang, text in translation_dict.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__): translation_tuples.append((lang, text)) else: translation_tuples.extend([(lang, el) for el in text]) # Ensure translations are in ascending order by language code. SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = zip(*sorted(lowerCAmelCase__)) return {"language": languages, "translation": translations} def _SCREAMING_SNAKE_CASE ( self : List[str]): from .features import Sequence, Value return { "language": Sequence(Value("string")), "translation": Sequence(Value("string")), }
13
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer lowerCAmelCase : Any = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast lowerCAmelCase : Dict = TaTokenizerFast lowerCAmelCase : Optional[int] = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : int = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Tuple = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys lowerCAmelCase : Optional[Any] = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
13
1
def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :int ) -> int: __lowerCAmelCase : Any = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __lowerCAmelCase : Dict = n - k # Calculate C(n,k) for i in range(SCREAMING_SNAKE_CASE ): result *= n - i result //= i + 1 return result def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int ) -> int: return binomial_coefficient(2 * node_count , SCREAMING_SNAKE_CASE ) // (node_count + 1) def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int ) -> int: if n < 0: raise ValueError("""factorial() not defined for negative values""" ) __lowerCAmelCase : Optional[int] = 1 for i in range(1 , n + 1 ): result *= i return result def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int ) -> int: return catalan_number(SCREAMING_SNAKE_CASE ) * factorial(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": _UpperCAmelCase = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
232
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { 'EleutherAI/gpt-j-6B': 'https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json', # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class snake_case_ ( __lowercase ): A_ = 'gptj' A_ = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Tuple , _snake_case : Optional[int]=50400 , _snake_case : Union[str, Any]=2048 , _snake_case : Tuple=4096 , _snake_case : Optional[int]=28 , _snake_case : Tuple=16 , _snake_case : Optional[Any]=64 , _snake_case : Optional[int]=None , _snake_case : str="gelu_new" , _snake_case : str=0.0 , _snake_case : Optional[int]=0.0 , _snake_case : List[str]=0.0 , _snake_case : Tuple=1E-5 , _snake_case : List[str]=0.02 , _snake_case : Optional[int]=True , _snake_case : Optional[Any]=50256 , _snake_case : List[str]=50256 , _snake_case : str=False , **_snake_case : Tuple , )->List[Any]: '''simple docstring''' __lowerCAmelCase : str = vocab_size __lowerCAmelCase : Any = n_positions __lowerCAmelCase : Optional[Any] = n_embd __lowerCAmelCase : Optional[int] = n_layer __lowerCAmelCase : Optional[int] = n_head __lowerCAmelCase : List[Any] = n_inner __lowerCAmelCase : List[str] = rotary_dim __lowerCAmelCase : int = activation_function __lowerCAmelCase : str = resid_pdrop __lowerCAmelCase : Union[str, Any] = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[Any] = layer_norm_epsilon __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : Dict = use_cache __lowerCAmelCase : List[str] = bos_token_id __lowerCAmelCase : List[Any] = eos_token_id super().__init__( bos_token_id=_snake_case , eos_token_id=_snake_case , tie_word_embeddings=_snake_case , **_snake_case ) class snake_case_ ( __lowercase ): def __init__( self : Union[str, Any] , _snake_case : PretrainedConfig , _snake_case : str = "default" , _snake_case : List[PatchingSpec] = None , _snake_case : bool = False , )->str: '''simple docstring''' super().__init__(_snake_case , task=_snake_case , patching_specs=_snake_case , use_past=_snake_case ) if not getattr(self._config , """pad_token_id""" , _snake_case ): # TODO: how to do that better? __lowerCAmelCase : Dict = 0 @property def UpperCAmelCase__ ( self : Tuple )->Mapping[str, Mapping[int, str]]: '''simple docstring''' __lowerCAmelCase : Any = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(_snake_case , direction="""inputs""" ) __lowerCAmelCase : Dict = {0: """batch""", 1: """past_sequence + sequence"""} else: __lowerCAmelCase : Optional[Any] = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCAmelCase__ ( self : Dict )->int: '''simple docstring''' return self._config.n_layer @property def UpperCAmelCase__ ( self : str )->int: '''simple docstring''' return self._config.n_head def UpperCAmelCase__ ( self : str , _snake_case : PreTrainedTokenizer , _snake_case : int = -1 , _snake_case : int = -1 , _snake_case : bool = False , _snake_case : Optional[TensorType] = None , )->Mapping[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = super(_snake_case , self ).generate_dummy_inputs( _snake_case , batch_size=_snake_case , seq_length=_snake_case , is_pair=_snake_case , framework=_snake_case ) # We need to order the input in the way they appears in the forward() __lowerCAmelCase : Union[str, Any] = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __lowerCAmelCase , __lowerCAmelCase : List[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values __lowerCAmelCase : Optional[Any] = seqlen + 2 __lowerCAmelCase : Dict = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __lowerCAmelCase : Tuple = [ (torch.zeros(_snake_case ), torch.zeros(_snake_case )) for _ in range(self.num_layers ) ] __lowerCAmelCase : int = common_inputs["""attention_mask"""] if self.use_past: __lowerCAmelCase : Union[str, Any] = ordered_inputs["""attention_mask"""].dtype __lowerCAmelCase : List[str] = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(_snake_case , _snake_case , dtype=_snake_case )] , dim=1 ) return ordered_inputs @property def UpperCAmelCase__ ( self : str )->int: '''simple docstring''' return 13
232
1
"""simple docstring""" import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = {"vocab_file": "spiece.model"} SCREAMING_SNAKE_CASE__ = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } SCREAMING_SNAKE_CASE__ = { "AI-Sweden/gpt-sw3-126m": 2_048, "AI-Sweden/gpt-sw3-350m": 2_048, "AI-Sweden/gpt-sw3-1.6b": 2_048, "AI-Sweden/gpt-sw3-6.7b": 2_048, "AI-Sweden/gpt-sw3-20b": 2_048, } class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase : int = VOCAB_FILES_NAMES _lowerCAmelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase : Dict = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase = None , **lowerCAmelCase , ): """simple docstring""" snake_case = {} if sp_model_kwargs is None else sp_model_kwargs snake_case = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) snake_case = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing snake_case = '<|endoftext|>' if eos_token is None else eos_token snake_case = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: snake_case = unk_token if pad_token is None else pad_token snake_case = eos_token if bos_token is None else bos_token else: snake_case = '<pad>' if pad_token is None else pad_token snake_case = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=lowerCAmelCase , remove_space=lowerCAmelCase , keep_accents=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , pad_token=lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase , ) snake_case = do_lower_case snake_case = remove_space snake_case = keep_accents snake_case = vocab_file snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase ) # Used for whitespace normalization in input texts # fmt : off snake_case = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing snake_case = re.compile( F"""[{"".join(map(lowerCAmelCase , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(1_27 , 1_60 ) ) + [1_60, 1_73, 82_03] ) )}]""" ) def __getstate__( self ): """simple docstring""" snake_case = self.__dict__.copy() snake_case = None return state def __setstate__( self , lowerCAmelCase ): """simple docstring""" snake_case = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): snake_case = {} snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def snake_case ( self ): """simple docstring""" return len(self.sp_model ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = self.non_printing_characters_re.sub('' , lowerCAmelCase ) # Normalize whitespaces snake_case = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization snake_case = unicodedata.normalize('NFC' , lowerCAmelCase ) return text def snake_case ( self , lowerCAmelCase , **lowerCAmelCase ): """simple docstring""" snake_case = self.preprocess_text(lowerCAmelCase ) return self.sp_model.encode(lowerCAmelCase , out_type=lowerCAmelCase ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" return self.sp_model.PieceToId(lowerCAmelCase ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" return self.sp_model.IdToPiece(lowerCAmelCase ) @staticmethod def snake_case ( lowerCAmelCase ): """simple docstring""" return out_string def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = [] snake_case = '' snake_case = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCAmelCase ) + token snake_case = True snake_case = [] else: current_sub_tokens.append(lowerCAmelCase ) snake_case = False out_string += self.sp_model.decode(lowerCAmelCase ) return out_string def snake_case ( self ): """simple docstring""" snake_case = {self.convert_ids_to_tokens(lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def snake_case ( self , lowerCAmelCase , lowerCAmelCase = None ): """simple docstring""" if not os.path.isdir(lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case = os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase , 'wb' ) as fi: snake_case = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase ) return (out_vocab_file,) def snake_case ( self , lowerCAmelCase , lowerCAmelCase = False ): """simple docstring""" if isinstance(lowerCAmelCase , lowerCAmelCase ): snake_case = self.preprocess_text(lowerCAmelCase ) snake_case = self.sp_model.encode(lowerCAmelCase ) else: snake_case = [self.preprocess_text(lowerCAmelCase ) for t in text] snake_case = self.sp_model.encode(lowerCAmelCase ) if return_tensors is True or return_tensors == "pt": snake_case = torch.tensor(lowerCAmelCase ) return token_ids def snake_case ( self , lowerCAmelCase ): """simple docstring""" return self.sp_model.decode(lowerCAmelCase ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = [F"""User: {text}""" if is_user else F"""Bot: {text}""" for is_user, text in conversation.iter_texts()] snake_case = ( F"""{self.eos_token}{self.bos_token}""" + F"""{self.bos_token}""".join(lowerCAmelCase ) + F"""{self.bos_token}Bot:""" ) return self.encode(text=lowerCAmelCase )
150
"""simple docstring""" import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient SCREAMING_SNAKE_CASE__ = WebClient(token=os.environ["CI_SLACK_BOT_TOKEN"]) def lowerCAmelCase__ ( _UpperCamelCase : Tuple ) -> List[Any]: """simple docstring""" snake_case = test_results.split(' ' ) snake_case = 0 snake_case = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. snake_case = expressions[-2] if '=' in expressions[-1] else expressions[-1] for i, expression in enumerate(_UpperCamelCase ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def lowerCAmelCase__ ( _UpperCamelCase : List[Any] ) -> List[str]: """simple docstring""" snake_case = {} snake_case = None snake_case = False for line in failures_short_lines.split('\n' ): if re.search(r'_ \[doctest\]' , _UpperCamelCase ): snake_case = True snake_case = line.split(' ' )[2] elif in_error and not line.split(' ' )[0].isdigit(): snake_case = line snake_case = False return failures class lowerCAmelCase_ : """simple docstring""" def __init__( self , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" snake_case = title snake_case = doc_test_results['time_spent'].split(',' )[0] snake_case = doc_test_results['success'] snake_case = doc_test_results['failures'] snake_case = self.n_success + self.n_failures # Failures and success of the modeling tests snake_case = doc_test_results @property def snake_case ( self ): """simple docstring""" snake_case = [self._time_spent] snake_case = 0 for time in time_spent: snake_case = time.split(':' ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(lowerCAmelCase ) == 1: snake_case = [0, 0, time_parts[0]] snake_case ,snake_case ,snake_case = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 36_00 + minutes * 60 + seconds snake_case ,snake_case ,snake_case = total_secs // 36_00, (total_secs % 36_00) // 60, total_secs % 60 return F"""{int(lowerCAmelCase )}h{int(lowerCAmelCase )}m{int(lowerCAmelCase )}s""" @property def snake_case ( self ): """simple docstring""" return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def snake_case ( self ): """simple docstring""" return { "type": "section", "text": { "type": "plain_text", "text": F"""🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.""", "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } @property def snake_case ( self ): """simple docstring""" return { "type": "section", "text": { "type": "plain_text", "text": ( F"""There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in""" F""" {self.time}.""" ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } @property def snake_case ( self ): """simple docstring""" snake_case = 40 snake_case = {k: v['failed'] for k, v in doc_test_results.items() if isinstance(lowerCAmelCase , lowerCAmelCase )} snake_case = '' for category, failures in category_failures.items(): if len(lowerCAmelCase ) == 0: continue if report != "": report += "\n\n" report += F"""*{category} failures*:""".ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(lowerCAmelCase ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": F"""The following examples had failures:\n\n\n{report}\n""", }, } @property def snake_case ( self ): """simple docstring""" snake_case = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(lowerCAmelCase ) @staticmethod def snake_case ( ): """simple docstring""" snake_case = [ { 'type': 'section', 'text': { 'type': 'plain_text', 'text': 'There was an issue running the tests.', }, 'accessory': { 'type': 'button', 'text': {'type': 'plain_text', 'text': 'Check Action results', 'emoji': True}, 'url': F"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } ] print('Sending the following payload' ) print(json.dumps({'blocks': json.loads(lowerCAmelCase )} ) ) client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , text='There was an issue running the tests.' , blocks=lowerCAmelCase , ) def snake_case ( self ): """simple docstring""" print('Sending the following payload' ) print(json.dumps({'blocks': json.loads(self.payload )} ) ) snake_case = F"""{self.n_failures} failures out of {self.n_tests} tests,""" if self.n_failures else 'All tests passed.' snake_case = client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , blocks=self.payload , text=lowerCAmelCase , ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" snake_case = '' for key, value in failures.items(): snake_case = value[:2_00] + ' [Truncated]' if len(lowerCAmelCase ) > 2_50 else value failures_text += F"""*{key}*\n_{value}_\n\n""" snake_case = job_name snake_case = {'type': 'section', 'text': {'type': 'mrkdwn', 'text': text}} if job_link is not None: snake_case = { 'type': 'button', 'text': {'type': 'plain_text', 'text': 'GitHub Action job', 'emoji': True}, 'url': job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def snake_case ( self ): """simple docstring""" if self.thread_ts is None: raise ValueError('Can only post reply if a post has been made.' ) snake_case = self.doc_test_results.pop('job_link' ) self.doc_test_results.pop('failures' ) self.doc_test_results.pop('success' ) self.doc_test_results.pop('time_spent' ) snake_case = sorted(self.doc_test_results.items() , key=lambda lowerCAmelCase : t[0] ) for job, job_result in sorted_dict: if len(job_result['failures'] ): snake_case = F"""*Num failures* :{len(job_result["failed"] )} \n""" snake_case = job_result['failures'] snake_case = self.get_reply_blocks(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , text=lowerCAmelCase ) print('Sending the following reply' ) print(json.dumps({'blocks': blocks} ) ) client.chat_postMessage( channel=os.environ['CI_SLACK_CHANNEL_ID_DAILY'] , text=F"""Results for {job}""" , blocks=lowerCAmelCase , thread_ts=self.thread_ts['ts'] , ) time.sleep(1 ) def lowerCAmelCase__ ( ) -> Tuple: """simple docstring""" snake_case = os.environ['GITHUB_RUN_ID'] snake_case = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100""" snake_case = requests.get(_UpperCamelCase ).json() snake_case = {} try: jobs.update({job['name']: job['html_url'] for job in result['jobs']} ) snake_case = math.ceil((result['total_count'] - 1_0_0) / 1_0_0 ) for i in range(_UpperCamelCase ): snake_case = requests.get(url + f"""&page={i + 2}""" ).json() jobs.update({job['name']: job['html_url'] for job in result['jobs']} ) return jobs except Exception as e: print('Unknown error, could not fetch links.' , _UpperCamelCase ) return {} def lowerCAmelCase__ ( _UpperCamelCase : str ) -> List[str]: """simple docstring""" snake_case = {} if os.path.exists(_UpperCamelCase ): snake_case = os.listdir(_UpperCamelCase ) for file in files: try: with open(os.path.join(_UpperCamelCase , _UpperCamelCase ) , encoding='utf-8' ) as f: snake_case = f.read() except UnicodeDecodeError as e: raise ValueError(f"""Could not open {os.path.join(_UpperCamelCase , _UpperCamelCase )}.""" ) from e return _artifact def lowerCAmelCase__ ( ) -> Union[str, Any]: """simple docstring""" class lowerCAmelCase_ : """simple docstring""" def __init__( self , lowerCAmelCase ): """simple docstring""" snake_case = name snake_case = [] def __str__( self ): """simple docstring""" return self.name def snake_case ( self , lowerCAmelCase ): """simple docstring""" self.paths.append({'name': self.name, 'path': path} ) snake_case = {} snake_case = filter(os.path.isdir , os.listdir() ) for directory in directories: snake_case = directory if artifact_name not in _available_artifacts: snake_case = Artifact(_UpperCamelCase ) _available_artifacts[artifact_name].add_path(_UpperCamelCase ) return _available_artifacts if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = get_job_links() SCREAMING_SNAKE_CASE__ = retrieve_available_artifacts() SCREAMING_SNAKE_CASE__ = collections.OrderedDict( [ ("*.py", "API Examples"), ("*.md", "MD Examples"), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' SCREAMING_SNAKE_CASE__ = { v: { "failed": [], "failures": {}, } for v in docs.values() } # Link to the GitHub Action job SCREAMING_SNAKE_CASE__ = github_actions_job_links.get("run_doctests") SCREAMING_SNAKE_CASE__ = available_artifacts["doc_tests_gpu_test_reports"].paths[0] SCREAMING_SNAKE_CASE__ = retrieve_artifact(artifact_path["name"]) if "stats" in artifact: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = handle_test_results(artifact["stats"]) SCREAMING_SNAKE_CASE__ = failed SCREAMING_SNAKE_CASE__ = success SCREAMING_SNAKE_CASE__ = time_spent[1:-1] + ", " SCREAMING_SNAKE_CASE__ = extract_first_line_failure(artifact["failures_short"]) for line in artifact["summary_short"].split("\n"): if re.search("FAILED", line): SCREAMING_SNAKE_CASE__ = line.replace("FAILED ", "") SCREAMING_SNAKE_CASE__ = line.split()[0].replace("\n", "") if "::" in line: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = line.split("::") else: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): SCREAMING_SNAKE_CASE__ = docs[file_regex] doc_test_results[category]["failed"].append(test) SCREAMING_SNAKE_CASE__ = all_failures[test] if test in all_failures else "N/A" SCREAMING_SNAKE_CASE__ = failure break SCREAMING_SNAKE_CASE__ = Message("🤗 Results of the doc tests.", doc_test_results) message.post() message.post_reply()
150
1
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = checkpoints.load_tax_checkpoint(SCREAMING_SNAKE_CASE ) lowercase__ = flatten_dict(SCREAMING_SNAKE_CASE ) return flax_params def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = {} lowercase__ = { '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } lowercase__ = { '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowercase__ = '''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowercase__ = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowercase__ = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowercase__ = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , SCREAMING_SNAKE_CASE ) lowercase__ = new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowercase__ = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , SCREAMING_SNAKE_CASE ) lowercase__ = flax_dict[key] lowercase__ = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowercase__ = torch.from_numpy(converted_dict[key].T ) else: lowercase__ = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False ): """simple docstring""" lowercase__ = get_flax_param(SCREAMING_SNAKE_CASE ) if not use_large: lowercase__ = PixaStructVisionConfig() lowercase__ = PixaStructTextConfig() else: lowercase__ = PixaStructVisionConfig( hidden_size=15_36 , d_ff=39_68 , num_attention_heads=24 , num_hidden_layers=18 ) lowercase__ = PixaStructTextConfig(hidden_size=15_36 , d_ff=39_68 , num_heads=24 , num_layers=18 ) lowercase__ = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=SCREAMING_SNAKE_CASE ) lowercase__ = PixaStructForConditionalGeneration(SCREAMING_SNAKE_CASE ) lowercase__ = rename_and_convert_flax_params(SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) lowercase__ = AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) lowercase__ = PixaStructImageProcessor() lowercase__ = PixaStructProcessor(image_processor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) if use_large: lowercase__ = 40_96 lowercase__ = True # mkdir if needed os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) model.save_pretrained(SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) print('''Model saved in {}'''.format(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--t5x_checkpoint_path', default=None, type=str, help='Path to the original T5x checkpoint.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--use_large', action='store_true', help='Use large model.') parser.add_argument('--is_vqa', action='store_true', help='Use large model.') lowerCAmelCase = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
352
lowerCAmelCase = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_0000)] def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_00_00] number //= 10_00_00 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase = [None] * 1000_0000 lowerCAmelCase = True lowerCAmelCase = False def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore lowercase__ = chain(next_number(SCREAMING_SNAKE_CASE ) ) lowercase__ = number_chain while number < 10_00_00_00: lowercase__ = number_chain number *= 10 return number_chain def _a ( SCREAMING_SNAKE_CASE = 10_00_00_00 ): """simple docstring""" for i in range(1 , SCREAMING_SNAKE_CASE ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
93
0
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class UpperCAmelCase_ ( snake_case ): @staticmethod @abstractmethod def _lowerCamelCase ( UpperCamelCase_ ) -> str: raise NotImplementedError() @abstractmethod def _lowerCamelCase ( self ) -> Optional[int]: raise NotImplementedError()
249
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class UpperCAmelCase_ ( snake_case ): @staticmethod @abstractmethod def _lowerCamelCase ( UpperCamelCase_ ) -> Union[str, Any]: raise NotImplementedError() @abstractmethod def _lowerCamelCase ( self ) -> str: raise NotImplementedError()
249
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Dict = {'vocab_file': 'spm_char.model'} __snake_case : int = { 'vocab_file': { 'microsoft/speecht5_asr': 'https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model', 'microsoft/speecht5_tts': 'https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model', 'microsoft/speecht5_vc': 'https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model', } } __snake_case : List[str] = { 'microsoft/speecht5_asr': 1_024, 'microsoft/speecht5_tts': 1_024, 'microsoft/speecht5_vc': 1_024, } class __UpperCAmelCase ( _UpperCamelCase ): '''simple docstring''' __lowercase : str = VOCAB_FILES_NAMES __lowercase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __lowercase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> None: A_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **_SCREAMING_SNAKE_CASE , ) A_ = vocab_file A_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_SCREAMING_SNAKE_CASE ) @property def __A ( self ) -> List[str]: return self.sp_model.get_piece_size() def __A ( self ) -> Tuple: A_ = {self.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> int: A_ = self.__dict__.copy() A_ = None return state def __setstate__( self , _SCREAMING_SNAKE_CASE ) -> str: A_ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): A_ = {} A_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __A ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: return self.sp_model.encode(_SCREAMING_SNAKE_CASE , out_type=_SCREAMING_SNAKE_CASE ) def __A ( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: return self.sp_model.piece_to_id(_SCREAMING_SNAKE_CASE ) def __A ( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: A_ = self.sp_model.IdToPiece(_SCREAMING_SNAKE_CASE ) return token def __A ( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: A_ = [] A_ = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_SCREAMING_SNAKE_CASE ) + token A_ = [] else: current_sub_tokens.append(_SCREAMING_SNAKE_CASE ) out_string += self.sp_model.decode(_SCREAMING_SNAKE_CASE ) return out_string.strip() def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_SCREAMING_SNAKE_CASE , token_ids_a=_SCREAMING_SNAKE_CASE , already_has_special_tokens=_SCREAMING_SNAKE_CASE ) A_ = [1] if token_ids_a is None: return ([0] * len(_SCREAMING_SNAKE_CASE )) + suffix_ones return ([0] * len(_SCREAMING_SNAKE_CASE )) + ([0] * len(_SCREAMING_SNAKE_CASE )) + suffix_ones def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A_ = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(_SCREAMING_SNAKE_CASE , '''wb''' ) as fi: A_ = self.sp_model.serialized_model_proto() fi.write(_SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
18
'''simple docstring''' from statistics import mean, stdev def _UpperCAmelCase ( _UpperCamelCase : list, _UpperCamelCase : int = 3 ) -> list: A_ = min(_UpperCamelCase ) A_ = max(_UpperCamelCase ) # normalize data return [round((x - x_min) / (x_max - x_min), _UpperCamelCase ) for x in data] def _UpperCAmelCase ( _UpperCamelCase : list, _UpperCamelCase : int = 3 ) -> list: A_ = mean(_UpperCamelCase ) A_ = stdev(_UpperCamelCase ) # standardize data return [round((x - mu) / (sigma), _UpperCamelCase ) for x in data]
18
1
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast lowerCAmelCase : Union[str, Any] = datasets.utils.logging.get_logger(__name__) @dataclass class _A ( datasets.BuilderConfig): SCREAMING_SNAKE_CASE : int = 10_000 SCREAMING_SNAKE_CASE : Optional[List[str]] = None SCREAMING_SNAKE_CASE : Optional[datasets.Features] = None class _A ( datasets.ArrowBasedBuilder): SCREAMING_SNAKE_CASE : Dict = ParquetConfig def UpperCAmelCase ( self ): """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE ): """simple docstring""" if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}" ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(_A , (str, list, tuple) ): SCREAMING_SNAKE_CASE_ : str = data_files if isinstance(_A , _A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive SCREAMING_SNAKE_CASE_ : List[str] = [dl_manager.iter_files(_A ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files} )] SCREAMING_SNAKE_CASE_ : Dict = [] for split_name, files in data_files.items(): if isinstance(_A , _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive SCREAMING_SNAKE_CASE_ : Dict = [dl_manager.iter_files(_A ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(_A ): with open(_A , 'rb' ) as f: SCREAMING_SNAKE_CASE_ : List[Any] = datasets.Features.from_arrow_schema(pq.read_schema(_A ) ) break splits.append(datasets.SplitGenerator(name=_A , gen_kwargs={'files': files} ) ) return splits def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE ): """simple docstring""" if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example SCREAMING_SNAKE_CASE_ : str = table_cast(_A , self.info.features.arrow_schema ) return pa_table def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( f"Tried to load parquet data with columns \'{self.config.columns}\' with mismatching features \'{self.info.features}\'" ) for file_idx, file in enumerate(itertools.chain.from_iterable(_A ) ): with open(_A , 'rb' ) as f: SCREAMING_SNAKE_CASE_ : Optional[int] = pq.ParquetFile(_A ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): SCREAMING_SNAKE_CASE_ : Optional[Any] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield f"{file_idx}_{batch_idx}", self._cast_table(_A ) except ValueError as e: logger.error(f"Failed to read file \'{file}\' with error {type(_A )}: {e}" ) raise
253
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase :Optional[int] = logging.get_logger(__name__) class _lowerCamelCase ( lowercase__ ): '''simple docstring''' A_ : Tuple = ["""pixel_values"""] def __init__( self : Dict , _A : bool = True , _A : Dict[str, int] = None , _A : float = None , _A : PILImageResampling = PILImageResampling.BILINEAR , _A : bool = True , _A : Union[int, float] = 1 / 255 , _A : bool = True , _A : Optional[Union[float, List[float]]] = None , _A : Optional[Union[float, List[float]]] = None , **_A : int , ) -> None: super().__init__(**_A ) __magic_name__ : List[str] = size if size is not None else {'shortest_edge': 384} __magic_name__ : Dict = get_size_dict(_A , default_to_square=_A ) __magic_name__ : List[Any] = do_resize __magic_name__ : str = size # Default value set here for backwards compatibility where the value in config is None __magic_name__ : Optional[Any] = crop_pct if crop_pct is not None else 224 / 256 __magic_name__ : int = resample __magic_name__ : List[str] = do_rescale __magic_name__ : List[Any] = rescale_factor __magic_name__ : str = do_normalize __magic_name__ : List[str] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __magic_name__ : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowerCAmelCase ( self : Optional[Any] , _A : np.ndarray , _A : Dict[str, int] , _A : float , _A : PILImageResampling = PILImageResampling.BICUBIC , _A : Optional[Union[str, ChannelDimension]] = None , **_A : Tuple , ) -> np.ndarray: __magic_name__ : Optional[int] = get_size_dict(_A , default_to_square=_A ) if "shortest_edge" not in size: raise ValueError(F'Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}' ) __magic_name__ : Dict = size['shortest_edge'] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct __magic_name__ : Dict = int(shortest_edge / crop_pct ) __magic_name__ : str = get_resize_output_image_size(_A , size=_A , default_to_square=_A ) __magic_name__ : Optional[int] = resize(image=_A , size=_A , resample=_A , data_format=_A , **_A ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=_A , size=(shortest_edge, shortest_edge) , data_format=_A , **_A ) else: # warping (no cropping) when evaluated at 384 or larger return resize( _A , size=(shortest_edge, shortest_edge) , resample=_A , data_format=_A , **_A ) def __lowerCAmelCase ( self : int , _A : np.ndarray , _A : Union[int, float] , _A : Optional[Union[str, ChannelDimension]] = None , **_A : int , ) -> int: return rescale(_A , scale=_A , data_format=_A , **_A ) def __lowerCAmelCase ( self : List[Any] , _A : np.ndarray , _A : Union[float, List[float]] , _A : Union[float, List[float]] , _A : Optional[Union[str, ChannelDimension]] = None , **_A : int , ) -> np.ndarray: return normalize(_A , mean=_A , std=_A , data_format=_A , **_A ) def __lowerCAmelCase ( self : Optional[Any] , _A : ImageInput , _A : bool = None , _A : Dict[str, int] = None , _A : float = None , _A : PILImageResampling = None , _A : bool = None , _A : float = None , _A : bool = None , _A : Optional[Union[float, List[float]]] = None , _A : Optional[Union[float, List[float]]] = None , _A : Optional[Union[str, TensorType]] = None , _A : ChannelDimension = ChannelDimension.FIRST , **_A : str , ) -> PIL.Image.Image: __magic_name__ : int = do_resize if do_resize is not None else self.do_resize __magic_name__ : Optional[int] = crop_pct if crop_pct is not None else self.crop_pct __magic_name__ : Optional[Any] = resample if resample is not None else self.resample __magic_name__ : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale __magic_name__ : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor __magic_name__ : str = do_normalize if do_normalize is not None else self.do_normalize __magic_name__ : str = image_mean if image_mean is not None else self.image_mean __magic_name__ : Dict = image_std if image_std is not None else self.image_std __magic_name__ : Dict = size if size is not None else self.size __magic_name__ : List[Any] = get_size_dict(_A , default_to_square=_A ) __magic_name__ : int = make_list_of_images(_A ) if not valid_images(_A ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError('crop_pct must be specified if size < 384.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __magic_name__ : Optional[Any] = [to_numpy_array(_A ) for image in images] if do_resize: __magic_name__ : List[str] = [self.resize(image=_A , size=_A , crop_pct=_A , resample=_A ) for image in images] if do_rescale: __magic_name__ : Tuple = [self.rescale(image=_A , scale=_A ) for image in images] if do_normalize: __magic_name__ : int = [self.normalize(image=_A , mean=_A , std=_A ) for image in images] __magic_name__ : Tuple = [to_channel_dimension_format(_A , _A ) for image in images] __magic_name__ : Union[str, Any] = {'pixel_values': images} return BatchFeature(data=_A , tensor_type=_A )
331
0
'''simple docstring''' snake_case__ = [ "VerificationMode", "Version", "disable_progress_bar", "enable_progress_bar", "is_progress_bar_enabled", "experimental", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
365
'''simple docstring''' import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class UpperCamelCase_ : """simple docstring""" def __init__( self : Optional[Any] , _lowerCamelCase : int , _lowerCamelCase : List[str]=3 , _lowerCamelCase : Any=32 , _lowerCamelCase : Union[str, Any]=3 , _lowerCamelCase : int=10 , _lowerCamelCase : Union[str, Any]=[8, 16, 32, 64] , _lowerCamelCase : Dict=[1, 1, 2, 1] , _lowerCamelCase : Union[str, Any]=True , _lowerCamelCase : Optional[int]=True , _lowerCamelCase : Any="relu" , _lowerCamelCase : Optional[Any]=3 , _lowerCamelCase : Optional[Any]=None , _lowerCamelCase : Dict=["stage2", "stage3", "stage4"] , _lowerCamelCase : Union[str, Any]=[2, 3, 4] , _lowerCamelCase : Tuple=1 , ): """simple docstring""" A_ : List[str] = parent A_ : List[str] = batch_size A_ : Union[str, Any] = image_size A_ : Tuple = num_channels A_ : Any = embeddings_size A_ : int = hidden_sizes A_ : Optional[Any] = depths A_ : List[Any] = is_training A_ : Optional[int] = use_labels A_ : int = hidden_act A_ : Tuple = num_labels A_ : Union[str, Any] = scope A_ : List[Any] = len(_lowerCamelCase ) A_ : Union[str, Any] = out_features A_ : List[Any] = out_indices A_ : Dict = num_groups def _a ( self : Optional[int] ): """simple docstring""" A_ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Union[str, Any] = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.num_labels ) A_ : Any = self.get_config() return config, pixel_values, labels def _a ( self : Union[str, Any] ): """simple docstring""" return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def _a ( self : List[Any] , _lowerCamelCase : List[str] , _lowerCamelCase : List[str] , _lowerCamelCase : Optional[Any] ): """simple docstring""" A_ : Any = BitModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCamelCase : List[Any] , _lowerCamelCase : str , _lowerCamelCase : Optional[int] ): """simple docstring""" A_ : Dict = self.num_labels A_ : Optional[Any] = BitForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : List[Any] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Any , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : List[Any] ): """simple docstring""" A_ : List[Any] = BitBackbone(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None A_ : Optional[Any] = None A_ : int = BitBackbone(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = model(_lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[Any] ): """simple docstring""" A_ : Union[str, Any] = self.prepare_config_and_inputs() A_ ,A_ ,A_ : Union[str, Any] = config_and_inputs A_ : str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase_ (a__, a__, unittest.TestCase ): """simple docstring""" _lowerCAmelCase = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () _lowerCAmelCase = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def _a ( self : Optional[Any] ): """simple docstring""" A_ : List[str] = BitModelTester(self ) A_ : Optional[Any] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def _a ( self : Optional[Any] ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ): """simple docstring""" return @unittest.skip(reason='''Bit does not output attentions''' ) def _a ( self : str ): """simple docstring""" pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def _a ( self : Union[str, Any] ): """simple docstring""" pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def _a ( self : Any ): """simple docstring""" pass def _a ( self : List[Any] ): """simple docstring""" A_ ,A_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Dict = model_class(_lowerCamelCase ) A_ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Union[str, Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def _a ( self : Optional[Any] ): """simple docstring""" A_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def _a ( self : Optional[Any] ): """simple docstring""" A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCamelCase ) def _a ( self : Tuple ): """simple docstring""" A_ ,A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : str = model_class(config=_lowerCamelCase ) for name, module in model.named_modules(): if isinstance(_lowerCamelCase , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) def _a ( self : int ): """simple docstring""" def check_hidden_states_output(_lowerCamelCase : Union[str, Any] , _lowerCamelCase : Dict , _lowerCamelCase : int ): A_ : Union[str, Any] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): A_ : Union[str, Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) A_ : int = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : List[Any] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) A_ ,A_ : str = self.model_tester.prepare_config_and_inputs_for_common() A_ : Tuple = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Tuple = layer_type A_ : Optional[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : List[str] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def _a ( self : Tuple ): """simple docstring""" pass def _a ( self : str ): """simple docstring""" A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def _a ( self : Union[str, Any] ): """simple docstring""" for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : List[Any] = BitModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def snake_case__ ( ) -> Optional[int]: A_ : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class UpperCamelCase_ (unittest.TestCase ): """simple docstring""" @cached_property def _a ( self : List[Any] ): """simple docstring""" return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _a ( self : Dict ): """simple docstring""" A_ : Optional[int] = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_lowerCamelCase ) A_ : Union[str, Any] = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : int = image_processor(images=_lowerCamelCase , return_tensors='''pt''' ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Union[str, Any] = model(**_lowerCamelCase ) # verify the logits A_ : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Tuple = torch.tensor([[-0.65_26, -0.52_63, -1.43_98]] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1E-4 ) ) @require_torch class UpperCamelCase_ (a__, unittest.TestCase ): """simple docstring""" _lowerCAmelCase = (BitBackbone,) if is_torch_available() else () _lowerCAmelCase = BitConfig _lowerCAmelCase = False def _a ( self : List[str] ): """simple docstring""" A_ : Union[str, Any] = BitModelTester(self )
4
0