code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
def __magic_name__ ( __a : Any ): '''simple docstring''' UpperCamelCase__ = 1 UpperCamelCase__ = 2 while i * i <= n: UpperCamelCase__ = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ = 1 UpperCamelCase__ = 1 while True: i += 1 t_num += i if count_divisors(SCREAMING_SNAKE_CASE__ ) > 500: break return t_num if __name__ == "__main__": print(solution())
244
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = GPTSanJapaneseTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = {"""do_clean_text""": False, """add_prefix_space""": False} def UpperCAmelCase__ ( self : Union[str, Any] ): super().setUp() # fmt: off __snake_case: str = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __snake_case: List[Any] = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __snake_case: Optional[int] = {"""unk_token""": """<unk>"""} __snake_case: Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __snake_case: int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(A ) ) def UpperCAmelCase__ ( self : Optional[int] , **A : int ): kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **A ) def UpperCAmelCase__ ( self : Optional[Any] , A : Dict ): __snake_case: Tuple = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __snake_case: str = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase__ ( self : Optional[int] , A : Optional[int] ): __snake_case , __snake_case: Optional[int] = self.get_input_output_texts(A ) __snake_case: Optional[Any] = tokenizer.encode(A , add_special_tokens=A ) __snake_case: int = tokenizer.decode(A , clean_up_tokenization_spaces=A ) return text, ids def UpperCAmelCase__ ( self : int ): pass # TODO add if relevant def UpperCAmelCase__ ( self : Dict ): pass # TODO add if relevant def UpperCAmelCase__ ( self : Optional[int] ): pass # TODO add if relevant def UpperCAmelCase__ ( self : Union[str, Any] ): __snake_case: Dict = self.get_tokenizer() # Testing tokenization __snake_case: List[str] = """こんにちは、世界。 こんばんは、㔺界。""" __snake_case: Optional[int] = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __snake_case: Any = tokenizer.tokenize(A ) self.assertListEqual(A , A ) # Testing conversion to ids without special tokens __snake_case: str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __snake_case: str = tokenizer.convert_tokens_to_ids(A ) self.assertListEqual(A , A ) # Testing conversion to ids with special tokens __snake_case: Optional[int] = tokens + [tokenizer.unk_token] __snake_case: List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __snake_case: str = tokenizer.convert_tokens_to_ids(A ) self.assertListEqual(A , A ) def UpperCAmelCase__ ( self : str ): __snake_case: Union[str, Any] = self.get_tokenizer() # Testing tokenization __snake_case: Optional[Any] = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __snake_case: Optional[Any] = """こんにちは、、、、世界。こんばんは、、、、世界。""" __snake_case: Union[str, Any] = tokenizer.encode(A ) __snake_case: List[Any] = tokenizer.decode(A ) self.assertEqual(A , A ) @slow def UpperCAmelCase__ ( self : Any ): __snake_case: Dict = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __snake_case: Any = """こんにちは、世界。""" __snake_case: Tuple = """こんばんは、㔺界。😀""" __snake_case: Optional[Any] = """こんにちは、世界。こんばんは、世界。😀""" __snake_case: int = tokenizer.encode(prefix_text + input_text ) __snake_case: Union[str, Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __snake_case: Tuple = tokenizer.encode(A , prefix_text=A ) __snake_case: Union[str, Any] = tokenizer.decode(A ) __snake_case: Dict = tokenizer.decode(A ) __snake_case: Optional[Any] = tokenizer.decode(A ) self.assertEqual(A , A ) self.assertEqual(A , A ) self.assertEqual(A , A ) @slow def UpperCAmelCase__ ( self : List[str] ): __snake_case: int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __snake_case: Optional[int] = """こんにちは、世界。""" __snake_case: Any = """こんばんは、㔺界。😀""" __snake_case: Optional[int] = len(tokenizer.encode(A ) ) - 2 __snake_case: str = len(tokenizer.encode(A ) ) - 2 __snake_case: Dict = [1] + [0] * (len_prefix + len_text + 1) __snake_case: str = [1] * (len_prefix + len_text + 1) + [0] __snake_case: List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __snake_case: int = tokenizer(prefix_text + input_text ).token_type_ids __snake_case: Optional[int] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __snake_case: Tuple = tokenizer(A , prefix_text=A ).token_type_ids self.assertListEqual(A , A ) self.assertListEqual(A , A ) self.assertListEqual(A , A ) @slow def UpperCAmelCase__ ( self : Dict ): __snake_case: Dict = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __snake_case: int = tokenizer.encode("""あンいワ""" ) __snake_case: Optional[int] = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __snake_case: List[Any] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(A ) , tokenizer.decode(A ) ) self.assertEqual(tokenizer.decode(A ) , tokenizer.decode(A ) ) self.assertNotEqual(A , A ) self.assertNotEqual(A , A ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase__ ( self : int ): __snake_case: List[str] = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __snake_case: Union[str, Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __snake_case: Optional[int] = tokenizer(A , padding=A ) __snake_case: int = tokenizer.batch_encode_plus(A , padding=A ) # fmt: off __snake_case: List[str] = [[35_993, 8_640, 25_948, 35_998, 30_647, 35_675, 35_999, 35_999], [35_993, 10_382, 9_868, 35_998, 30_646, 9_459, 30_646, 35_675]] __snake_case: int = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __snake_case: Dict = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , A ) self.assertListEqual(x_token.token_type_ids , A ) self.assertListEqual(x_token.attention_mask , A ) self.assertListEqual(x_token_a.input_ids , A ) self.assertListEqual(x_token_a.token_type_ids , A ) self.assertListEqual(x_token_a.attention_mask , A ) def UpperCAmelCase__ ( self : Union[str, Any] ): # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase__ ( self : List[str] ): # tokenizer has no padding token pass
111
0
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class a_ : def __init__( self :Tuple , _lowercase :Dict , _lowercase :Union[str, Any]=13 , _lowercase :Optional[Any]=7 , _lowercase :List[Any]=True , _lowercase :Dict=True , _lowercase :Optional[Any]=True , _lowercase :str=True , _lowercase :str=99 , _lowercase :Optional[Any]=16 , _lowercase :List[Any]=36 , _lowercase :Dict=6 , _lowercase :Tuple=6 , _lowercase :Optional[int]=6 , _lowercase :int=37 , _lowercase :str="gelu" , _lowercase :List[Any]=0.1 , _lowercase :int=0.1 , _lowercase :Any=512 , _lowercase :str=16 , _lowercase :Optional[Any]=2 , _lowercase :str=0.02 , _lowercase :Tuple=3 , _lowercase :List[Any]=4 , _lowercase :str=None , ) -> List[str]: UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = embedding_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_hidden_groups UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_labels UpperCAmelCase_ = num_choices UpperCAmelCase_ = scope def __a ( self :int) -> Optional[Any]: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length]) UpperCAmelCase_ = None if self.use_token_type_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices) UpperCAmelCase_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __a ( self :int) -> Optional[Any]: return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def __a ( self :Optional[int] , _lowercase :Tuple , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Tuple , _lowercase :Optional[int] , _lowercase :Dict , _lowercase :List[Any]) -> List[Any]: UpperCAmelCase_ = AlbertModel(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase) UpperCAmelCase_ = model(_lowercase , token_type_ids=_lowercase) UpperCAmelCase_ = model(_lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def __a ( self :Any , _lowercase :str , _lowercase :str , _lowercase :str , _lowercase :List[str] , _lowercase :str , _lowercase :Any , _lowercase :Optional[int]) -> Any: UpperCAmelCase_ = AlbertForPreTraining(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase , sentence_order_label=_lowercase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def __a ( self :Union[str, Any] , _lowercase :List[Any] , _lowercase :Union[str, Any] , _lowercase :str , _lowercase :Dict , _lowercase :int , _lowercase :Tuple , _lowercase :Dict) -> Any: UpperCAmelCase_ = AlbertForMaskedLM(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def __a ( self :Dict , _lowercase :Union[str, Any] , _lowercase :Dict , _lowercase :List[Any] , _lowercase :List[str] , _lowercase :int , _lowercase :int , _lowercase :str) -> str: UpperCAmelCase_ = AlbertForQuestionAnswering(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , start_positions=_lowercase , end_positions=_lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def __a ( self :Tuple , _lowercase :List[str] , _lowercase :Any , _lowercase :Dict , _lowercase :List[Any] , _lowercase :Optional[Any] , _lowercase :Dict , _lowercase :Tuple) -> int: UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = AlbertForSequenceClassification(_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __a ( self :Union[str, Any] , _lowercase :Optional[int] , _lowercase :str , _lowercase :Optional[int] , _lowercase :int , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :List[str]) -> Any: UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = AlbertForTokenClassification(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def __a ( self :Union[str, Any] , _lowercase :Optional[Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Union[str, Any] , _lowercase :List[str] , _lowercase :int , _lowercase :Optional[Any]) -> int: UpperCAmelCase_ = self.num_choices UpperCAmelCase_ = AlbertForMultipleChoice(config=_lowercase) model.to(_lowercase) model.eval() UpperCAmelCase_ = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() UpperCAmelCase_ = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() UpperCAmelCase_ = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() UpperCAmelCase_ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def __a ( self :Any) -> Tuple: UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class a_ ( _snake_case , _snake_case , unittest.TestCase ): UpperCamelCase__ : Tuple =( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase__ : Tuple =( { "feature-extraction": AlbertModel, "fill-mask": AlbertForMaskedLM, "question-answering": AlbertForQuestionAnswering, "text-classification": AlbertForSequenceClassification, "token-classification": AlbertForTokenClassification, "zero-shot": AlbertForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase__ : List[str] =True def __a ( self :Optional[int] , _lowercase :str , _lowercase :Tuple , _lowercase :Dict=False) -> Dict: UpperCAmelCase_ = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase) if return_labels: if model_class in get_values(_lowercase): UpperCAmelCase_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_lowercase) UpperCAmelCase_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowercase) return inputs_dict def __a ( self :Any) -> List[str]: UpperCAmelCase_ = AlbertModelTester(self) UpperCAmelCase_ = ConfigTester(self , config_class=_lowercase , hidden_size=37) def __a ( self :int) -> Optional[Any]: self.config_tester.run_common_tests() def __a ( self :Optional[Any]) -> Optional[Any]: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase) def __a ( self :Any) -> List[str]: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_lowercase) def __a ( self :Dict) -> List[str]: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowercase) def __a ( self :Optional[int]) -> Tuple: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_lowercase) def __a ( self :str) -> Optional[int]: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowercase) def __a ( self :List[str]) -> Union[str, Any]: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowercase) def __a ( self :Optional[int]) -> int: UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase_ = type self.model_tester.create_and_check_model(*_lowercase) @slow def __a ( self :Optional[Any]) -> Dict: for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = AlbertModel.from_pretrained(_lowercase) self.assertIsNotNone(_lowercase) @require_torch class a_ ( unittest.TestCase ): @slow def __a ( self :str) -> List[str]: UpperCAmelCase_ = AlbertModel.from_pretrained('''albert-base-v2''') UpperCAmelCase_ = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]]) UpperCAmelCase_ = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): UpperCAmelCase_ = model(_lowercase , attention_mask=_lowercase)[0] UpperCAmelCase_ = torch.Size((1, 11, 768)) self.assertEqual(output.shape , _lowercase) UpperCAmelCase_ = torch.tensor( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowercase , atol=1E-4))
350
def A ( __UpperCAmelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase_ = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __UpperCAmelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
344
0
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class a__ ( tf.keras.layers.Layer ): def __init__( self : Tuple , a : Any , a : Any , a : Optional[Any] , a : Optional[int] , a : str=1 , a : Optional[int]=False , **a : str ): """simple docstring""" super().__init__(**a ) __lowerCamelCase = vocab_size __lowerCamelCase = d_embed __lowerCamelCase = d_proj __lowerCamelCase = cutoffs + [vocab_size] __lowerCamelCase = [0] + self.cutoffs __lowerCamelCase = div_val __lowerCamelCase = self.cutoffs[0] __lowerCamelCase = len(self.cutoffs ) - 1 __lowerCamelCase = self.shortlist_size + self.n_clusters __lowerCamelCase = keep_order __lowerCamelCase = [] __lowerCamelCase = [] def SCREAMING_SNAKE_CASE__ ( self : int , a : int ): """simple docstring""" if self.n_clusters > 0: __lowerCamelCase = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='''zeros''' , trainable=a , name='''cluster_weight''' ) __lowerCamelCase = self.add_weight( shape=(self.n_clusters,) , initializer='''zeros''' , trainable=a , name='''cluster_bias''' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: __lowerCamelCase = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='''zeros''' , trainable=a , name=f"""out_projs_._{i}""" , ) self.out_projs.append(a ) else: self.out_projs.append(a ) __lowerCamelCase = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='''zeros''' , trainable=a , name=f"""out_layers_._{i}_._weight""" , ) __lowerCamelCase = self.add_weight( shape=(self.vocab_size,) , initializer='''zeros''' , trainable=a , name=f"""out_layers_._{i}_._bias""" , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): __lowerCamelCase , __lowerCamelCase = self.cutoff_ends[i], self.cutoff_ends[i + 1] __lowerCamelCase = self.d_embed // (self.div_val**i) __lowerCamelCase = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='''zeros''' , trainable=a , name=f"""out_projs_._{i}""" ) self.out_projs.append(a ) __lowerCamelCase = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='''zeros''' , trainable=a , name=f"""out_layers_._{i}_._weight""" , ) __lowerCamelCase = self.add_weight( shape=(r_idx - l_idx,) , initializer='''zeros''' , trainable=a , name=f"""out_layers_._{i}_._bias""" , ) self.out_layers.append((weight, bias) ) super().build(a ) @staticmethod def SCREAMING_SNAKE_CASE__ ( a : Optional[Any] , a : Dict , a : Optional[Any] , a : List[str]=None ): """simple docstring""" __lowerCamelCase = x if proj is not None: __lowerCamelCase = tf.einsum('''ibd,ed->ibe''' , a , a ) return tf.einsum('''ibd,nd->ibn''' , a , a ) + b @staticmethod def SCREAMING_SNAKE_CASE__ ( a : List[str] , a : Tuple ): """simple docstring""" __lowerCamelCase = shape_list(a ) __lowerCamelCase = tf.range(lp_size[0] , dtype=target.dtype ) __lowerCamelCase = tf.stack([r, target] , 1 ) return tf.gather_nd(a , a ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : Optional[int] , a : Dict , a : int=True , a : Dict=False ): """simple docstring""" __lowerCamelCase = 0 if self.n_clusters == 0: __lowerCamelCase = self._logit(a , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: __lowerCamelCase = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=a , logits=a ) __lowerCamelCase = tf.nn.log_softmax(a , axis=-1 ) else: __lowerCamelCase = shape_list(a ) __lowerCamelCase = [] __lowerCamelCase = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): __lowerCamelCase , __lowerCamelCase = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: __lowerCamelCase = (target >= l_idx) & (target < r_idx) __lowerCamelCase = tf.where(a ) __lowerCamelCase = tf.boolean_mask(a , a ) - l_idx if self.div_val == 1: __lowerCamelCase = self.out_layers[0][0][l_idx:r_idx] __lowerCamelCase = self.out_layers[0][1][l_idx:r_idx] else: __lowerCamelCase = self.out_layers[i][0] __lowerCamelCase = self.out_layers[i][1] if i == 0: __lowerCamelCase = tf.concat([cur_W, self.cluster_weight] , 0 ) __lowerCamelCase = tf.concat([cur_b, self.cluster_bias] , 0 ) __lowerCamelCase = self._logit(a , a , a , self.out_projs[0] ) __lowerCamelCase = tf.nn.log_softmax(a ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: __lowerCamelCase = tf.boolean_mask(a , a ) __lowerCamelCase = self._gather_logprob(a , a ) else: __lowerCamelCase = self._logit(a , a , a , self.out_projs[i] ) __lowerCamelCase = tf.nn.log_softmax(a ) __lowerCamelCase = self.cutoffs[0] + i - 1 # No probability for the head cluster __lowerCamelCase = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(a ) if target is not None: __lowerCamelCase = tf.boolean_mask(a , a ) __lowerCamelCase = tf.boolean_mask(a , a ) __lowerCamelCase = self._gather_logprob(a , a ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(a , -cur_logprob , shape_list(a ) ) __lowerCamelCase = tf.concat(a , axis=-1 ) if target is not None: if return_mean: __lowerCamelCase = tf.reduce_mean(a ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(a ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(a , name=self.name , aggregation='''mean''' if return_mean else '''''' ) return out
67
'''simple docstring''' import pprint import requests UpperCamelCase__ = '''https://zenquotes.io/api''' def a__ ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/today''' ).json() def a__ ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/random''' ).json() if __name__ == "__main__": UpperCamelCase__ = random_quotes() pprint.pprint(response)
181
0
'''simple docstring''' import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def UpperCAmelCase_ ( *__lowerCamelCase : Union[str, Any] ): if not isinstance(__lowerCamelCase ,__lowerCamelCase ): lowercase_ :int = list(__lowerCamelCase ) for i in range(len(__lowerCamelCase ) ): lowercase_ :List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def UpperCAmelCase_ ( __lowerCamelCase : Exception ): lowercase_ :Any = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowerCamelCase ,__lowerCamelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def UpperCAmelCase_ ( __lowerCamelCase : callable = None ,__lowerCamelCase : int = 1_28 ): if function is None: return functools.partial(__lowerCamelCase ,starting_batch_size=__lowerCamelCase ) lowercase_ :Optional[Any] = starting_batch_size def decorator(*__lowerCamelCase : Any ,**__lowerCamelCase : int ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() lowercase_ :str = list(inspect.signature(__lowerCamelCase ).parameters.keys() ) # Guard against user error if len(__lowerCamelCase ) < (len(__lowerCamelCase ) + 1): lowercase_ :Optional[Any] = ", ".join([F'{arg}={value}' for arg, value in zip(params[1:] ,args[1:] )] ) raise TypeError( F'Batch size was passed into `{function.__name__}` as the first argument when called.' F'Remove this as the decorator already does so: `{function.__name__}({arg_str})`' ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowerCamelCase ,*__lowerCamelCase ,**__lowerCamelCase ) except Exception as e: if should_reduce_batch_size(__lowerCamelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
366
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : list ): if len(__lowerCamelCase ) <= 1: return lst lowercase_ :Optional[Any] = 1 while i < len(__lowerCamelCase ): if lst[i - 1] <= lst[i]: i += 1 else: lowercase_ , lowercase_ :int = lst[i], lst[i - 1] i -= 1 if i == 0: lowercase_ :Dict = 1 return lst if __name__ == "__main__": lowerCAmelCase : Any =input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase : List[str] =[int(item) for item in user_input.split(''',''')] print(gnome_sort(unsorted))
147
0
from timeit import timeit _UpperCAmelCase : Union[str, Any] = { """MALAYALAM""": True, """String""": False, """rotor""": True, """level""": True, """A""": True, """BB""": True, """ABC""": False, """amanaplanacanalpanama""": True, # "a man a plan a canal panama" } # Ensure our test data is valid assert all((key == key[::-1]) is value for key, value in test_data.items()) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: lowerCamelCase__ : Tuple = 0 lowerCamelCase__ : Dict = len(_UpperCAmelCase ) - 1 while start_i < end_i: if s[start_i] == s[end_i]: start_i += 1 end_i -= 1 else: return False return True def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: lowerCamelCase__ : Dict = len(_UpperCAmelCase ) // 2 lowerCamelCase__ : Tuple = len(_UpperCAmelCase ) # We need to traverse till half of the length of string # as we can get access of the i'th last element from # i'th index. # eg: [0,1,2,3,4,5] => 4th index can be accessed # with the help of 1st index (i==n-i-1) # where n is length of string return all(s[i] == s[n - i - 1] for i in range(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if len(_UpperCAmelCase ) <= 2: return True if s[0] == s[len(_UpperCAmelCase ) - 1]: return is_palindrome_recursive(s[1:-1] ) else: return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return s == s[::-1] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: lowerCamelCase__ : Any = F"""all({name}(key) is value for key, value in test_data.items())""" lowerCamelCase__ : List[Any] = F"""from __main__ import test_data, {name}""" lowerCamelCase__ : int = 50_0000 lowerCamelCase__ : List[str] = timeit(stmt=_UpperCAmelCase , setup=_UpperCAmelCase , number=_UpperCAmelCase ) print(F"""{name:<35} finished {number:,} runs in {result:.5f} seconds""" ) if __name__ == "__main__": for key, value in test_data.items(): assert is_palindrome(key) is is_palindrome_recursive(key) assert is_palindrome(key) is is_palindrome_slice(key) print(F"""{key:21} {value}""") print("""a man a plan a canal panama""") # finished 500,000 runs in 0.46793 seconds benchmark_function("""is_palindrome_slice""") # finished 500,000 runs in 0.85234 seconds benchmark_function("""is_palindrome""") # finished 500,000 runs in 1.32028 seconds benchmark_function("""is_palindrome_recursive""") # finished 500,000 runs in 2.08679 seconds benchmark_function("""is_palindrome_traversal""")
50
import argparse import os import re SCREAMING_SNAKE_CASE__ : List[Any] = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict SCREAMING_SNAKE_CASE__ : Optional[int] = re.compile(r'[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict') # re pattern that matches identifiers in mappings SCREAMING_SNAKE_CASE__ : Tuple = re.compile(r'\s*\(\s*"(\S[^"]+)"') def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = False ) -> int: with open(_SCREAMING_SNAKE_CASE ,"r" ,encoding="utf-8" ) as f: lowerCamelCase : List[Any] = f.read() lowerCamelCase : str = content.split("\n" ) lowerCamelCase : int = [] lowerCamelCase : List[Any] = 0 while line_idx < len(_SCREAMING_SNAKE_CASE ): if _re_intro_mapping.search(lines[line_idx] ) is not None: lowerCamelCase : Optional[int] = len(re.search(r"^(\s*)\S" ,lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(" " * indent + "(" ): new_lines.append(lines[line_idx] ) line_idx += 1 lowerCamelCase : Optional[int] = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": lowerCamelCase : List[str] = line_idx while not lines[line_idx].startswith(" " * indent + ")" ): line_idx += 1 blocks.append("\n".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers lowerCamelCase : Union[str, Any] = sorted(_SCREAMING_SNAKE_CASE ,key=lambda _SCREAMING_SNAKE_CASE : _re_identifier.search(_SCREAMING_SNAKE_CASE ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(_SCREAMING_SNAKE_CASE ,"w" ,encoding="utf-8" ) as f: f.write("\n".join(_SCREAMING_SNAKE_CASE ) ) elif "\n".join(_SCREAMING_SNAKE_CASE ) != content: return True def A ( _SCREAMING_SNAKE_CASE = False ) -> List[str]: lowerCamelCase : str = [os.path.join(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) for f in os.listdir(_SCREAMING_SNAKE_CASE ) if f.endswith(".py" )] lowerCamelCase : Union[str, Any] = [sort_auto_mapping(_SCREAMING_SNAKE_CASE ,overwrite=_SCREAMING_SNAKE_CASE ) for fname in fnames] if not overwrite and any(_SCREAMING_SNAKE_CASE ): lowerCamelCase : str = [f for f, d in zip(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) if d] raise ValueError( f'''The following files have auto mappings that need sorting: {", ".join(_SCREAMING_SNAKE_CASE )}. Run `make style` to fix''' " this." ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[str] = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') SCREAMING_SNAKE_CASE__ : List[str] = parser.parse_args() sort_all_auto_mappings(not args.check_only)
48
0
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class a : """simple docstring""" def __init__( self : str , snake_case : Union[str, Any] , snake_case : Union[str, Any]=2 , snake_case : Union[str, Any]=True , snake_case : List[Any]=False , snake_case : Tuple=10 , snake_case : Any=3 , snake_case : int=32 * 8 , snake_case : str=32 * 8 , snake_case : Dict=4 , snake_case : Union[str, Any]=64 , ) -> Tuple: __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Union[str, Any] = batch_size __UpperCAmelCase : Dict = is_training __UpperCAmelCase : Optional[int] = use_auxiliary_loss __UpperCAmelCase : List[str] = num_queries __UpperCAmelCase : Dict = num_channels __UpperCAmelCase : Union[str, Any] = min_size __UpperCAmelCase : List[Any] = max_size __UpperCAmelCase : Tuple = num_labels __UpperCAmelCase : str = hidden_dim __UpperCAmelCase : List[str] = hidden_dim def lowerCamelCase__ ( self : Optional[int] ) -> List[str]: __UpperCAmelCase : str = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( snake_case ) __UpperCAmelCase : Tuple = torch.ones([self.batch_size, self.min_size, self.max_size] , device=snake_case ) __UpperCAmelCase : Dict = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=snake_case ) > 0.5 ).float() __UpperCAmelCase : Dict = (torch.rand((self.batch_size, self.num_labels) , device=snake_case ) > 0.5).long() __UpperCAmelCase : Optional[int] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowerCamelCase__ ( self : int ) -> int: __UpperCAmelCase : Optional[int] = MaskaFormerConfig( hidden_size=self.hidden_dim , ) __UpperCAmelCase : Optional[Any] = self.num_queries __UpperCAmelCase : List[str] = self.num_labels __UpperCAmelCase : int = [1, 1, 1, 1] __UpperCAmelCase : Dict = self.num_channels __UpperCAmelCase : Tuple = 64 __UpperCAmelCase : int = 128 __UpperCAmelCase : Optional[Any] = self.hidden_dim __UpperCAmelCase : Optional[Any] = self.hidden_dim __UpperCAmelCase : int = self.hidden_dim return config def lowerCamelCase__ ( self : List[Any] ) -> List[str]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase : Union[str, Any] = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def lowerCamelCase__ ( self : Optional[int] , snake_case : Any , snake_case : Optional[Any] ) -> Tuple: __UpperCAmelCase : Union[str, Any] = output.encoder_hidden_states __UpperCAmelCase : Any = output.pixel_decoder_hidden_states __UpperCAmelCase : Union[str, Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(snake_case ) , config.decoder_layers ) def lowerCamelCase__ ( self : Optional[Any] , snake_case : Dict , snake_case : str , snake_case : Dict , snake_case : Any=False ) -> Optional[int]: with torch.no_grad(): __UpperCAmelCase : List[Any] = MaskaFormerModel(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : Optional[Any] = model(pixel_values=snake_case , pixel_mask=snake_case ) __UpperCAmelCase : Optional[Any] = model(snake_case , output_hidden_states=snake_case ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(snake_case , snake_case ) def lowerCamelCase__ ( self : Dict , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Optional[Any] , snake_case : Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase : int = MaskaFormerForUniversalSegmentation(config=snake_case ) model.to(snake_case ) model.eval() def comm_check_on_output(snake_case : List[Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): __UpperCAmelCase : int = model(pixel_values=snake_case , pixel_mask=snake_case ) __UpperCAmelCase : Any = model(snake_case ) comm_check_on_output(snake_case ) __UpperCAmelCase : List[str] = model( pixel_values=snake_case , pixel_mask=snake_case , mask_labels=snake_case , class_labels=snake_case ) comm_check_on_output(snake_case ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class a ( _a , _a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE : Union[str, Any] = {"feature-extraction": MaskaFormerModel} if is_torch_available() else {} SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Any = False SCREAMING_SNAKE_CASE : Dict = False def lowerCamelCase__ ( self : Optional[Any] ) -> Tuple: __UpperCAmelCase : Optional[int] = MaskaFormerModelTester(self ) __UpperCAmelCase : Optional[Any] = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case ) def lowerCamelCase__ ( self : Optional[Any] ) -> int: self.config_tester.run_common_tests() def lowerCamelCase__ ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(snake_case , **snake_case , output_hidden_states=snake_case ) def lowerCamelCase__ ( self : List[Any] ) -> Optional[int]: __UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*snake_case ) @unittest.skip(reason='''Mask2Former does not use inputs_embeds''' ) def lowerCamelCase__ ( self : List[Any] ) -> Optional[Any]: pass @unittest.skip(reason='''Mask2Former does not have a get_input_embeddings method''' ) def lowerCamelCase__ ( self : str ) -> List[Any]: pass @unittest.skip(reason='''Mask2Former is not a generative model''' ) def lowerCamelCase__ ( self : Optional[Any] ) -> List[Any]: pass @unittest.skip(reason='''Mask2Former does not use token embeddings''' ) def lowerCamelCase__ ( self : Optional[Any] ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip( reason='''Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def lowerCamelCase__ ( self : Union[str, Any] ) -> Dict: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def lowerCamelCase__ ( self : Union[str, Any] ) -> Optional[int]: pass def lowerCamelCase__ ( self : Tuple ) -> int: __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = model_class(snake_case ) __UpperCAmelCase : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Optional[int] = [*signature.parameters.keys()] __UpperCAmelCase : Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , snake_case ) @slow def lowerCamelCase__ ( self : Union[str, Any] ) -> Any: for model_name in ["facebook/mask2former-swin-small-coco-instance"]: __UpperCAmelCase : List[str] = MaskaFormerModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def lowerCamelCase__ ( self : List[str] ) -> int: __UpperCAmelCase : str = (self.model_tester.min_size,) * 2 __UpperCAmelCase : Tuple = { '''pixel_values''': torch.randn((2, 3, *size) , device=snake_case ), '''mask_labels''': torch.randn((2, 10, *size) , device=snake_case ), '''class_labels''': torch.zeros(2 , 10 , device=snake_case ).long(), } __UpperCAmelCase : Any = self.model_tester.get_config() __UpperCAmelCase : Union[str, Any] = MaskaFormerForUniversalSegmentation(snake_case ).to(snake_case ) __UpperCAmelCase : Any = model(**snake_case ) self.assertTrue(outputs.loss is not None ) def lowerCamelCase__ ( self : List[str] ) -> Optional[Any]: __UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(snake_case , **snake_case , output_hidden_states=snake_case ) def lowerCamelCase__ ( self : List[Any] ) -> Union[str, Any]: __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : Any = model_class(snake_case ).to(snake_case ) __UpperCAmelCase : str = model(**snake_case , output_attentions=snake_case ) self.assertTrue(outputs.attentions is not None ) def lowerCamelCase__ ( self : Tuple ) -> List[str]: if not self.model_tester.is_training: return __UpperCAmelCase : Optional[Any] = self.all_model_classes[1] __UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() __UpperCAmelCase : List[Any] = model_class(snake_case ) model.to(snake_case ) model.train() __UpperCAmelCase : Union[str, Any] = model(snake_case , mask_labels=snake_case , class_labels=snake_case ).loss loss.backward() def lowerCamelCase__ ( self : Optional[Any] ) -> str: __UpperCAmelCase : List[Any] = self.all_model_classes[1] __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : List[str] = model_class(snake_case ).to(snake_case ) model.train() __UpperCAmelCase : List[Any] = model(snake_case , mask_labels=snake_case , class_labels=snake_case ) __UpperCAmelCase : str = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __UpperCAmelCase : List[str] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() __UpperCAmelCase : Optional[int] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __UpperCAmelCase : Optional[Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=snake_case ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __UpperCAmelCase :List[str] = 1e-4 def _a ( ): '''simple docstring''' __UpperCAmelCase : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class a ( unittest.TestCase ): """simple docstring""" @cached_property def lowerCamelCase__ ( self : Any ) -> int: return "facebook/mask2former-swin-small-coco-instance" @cached_property def lowerCamelCase__ ( self : List[str] ) -> str: return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def lowerCamelCase__ ( self : Any ) -> Optional[int]: __UpperCAmelCase : Any = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(snake_case ) __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : Dict = prepare_img() __UpperCAmelCase : Dict = image_processor(snake_case , return_tensors='''pt''' ).to(snake_case ) __UpperCAmelCase : int = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(snake_case , (1, 3, 384, 384) ) with torch.no_grad(): __UpperCAmelCase : Tuple = model(**snake_case ) __UpperCAmelCase : Dict = torch.tensor( [[-0.2_790, -1.0_717, -1.1_668], [-0.5_128, -0.3_128, -0.4_987], [-0.5_832, 0.1_971, -0.0_197]] ).to(snake_case ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , snake_case , atol=snake_case ) ) __UpperCAmelCase : str = torch.tensor( [[0.8_973, 1.1_847, 1.1_776], [1.1_934, 1.5_040, 1.5_128], [1.1_153, 1.4_486, 1.4_951]] ).to(snake_case ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , snake_case , atol=snake_case ) ) __UpperCAmelCase : Tuple = torch.tensor( [[2.1_152, 1.7_000, -0.8_603], [1.5_808, 1.8_004, -0.9_353], [1.6_043, 1.7_495, -0.5_999]] ).to(snake_case ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , snake_case , atol=snake_case ) ) def lowerCamelCase__ ( self : Optional[int] ) -> int: __UpperCAmelCase : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(snake_case ).eval() __UpperCAmelCase : List[Any] = self.default_image_processor __UpperCAmelCase : Dict = prepare_img() __UpperCAmelCase : Dict = image_processor(snake_case , return_tensors='''pt''' ).to(snake_case ) __UpperCAmelCase : List[str] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(snake_case , (1, 3, 384, 384) ) with torch.no_grad(): __UpperCAmelCase : Optional[int] = model(**snake_case ) # masks_queries_logits __UpperCAmelCase : List[str] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) __UpperCAmelCase : Optional[Any] = [ [-8.7_839, -9.0_056, -8.8_121], [-7.4_104, -7.0_313, -6.5_401], [-6.6_105, -6.3_427, -6.4_675], ] __UpperCAmelCase : Dict = torch.tensor(snake_case ).to(snake_case ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , snake_case , atol=snake_case ) ) # class_queries_logits __UpperCAmelCase : Optional[Any] = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) __UpperCAmelCase : Dict = torch.tensor( [ [1.8_324, -8.0_835, -4.1_922], [0.8_450, -9.0_050, -3.6_053], [0.3_045, -7.7_293, -3.0_275], ] ).to(snake_case ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , snake_case , atol=snake_case ) ) def lowerCamelCase__ ( self : Optional[int] ) -> List[Any]: __UpperCAmelCase : Union[str, Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(snake_case ).eval() __UpperCAmelCase : Tuple = self.default_image_processor __UpperCAmelCase : Union[str, Any] = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors='''pt''' , ) __UpperCAmelCase : List[Any] = inputs['''pixel_values'''].to(snake_case ) __UpperCAmelCase : int = [el.to(snake_case ) for el in inputs['''mask_labels''']] __UpperCAmelCase : Optional[Any] = [el.to(snake_case ) for el in inputs['''class_labels''']] with torch.no_grad(): __UpperCAmelCase : Tuple = model(**snake_case ) self.assertTrue(outputs.loss is not None )
350
'''simple docstring''' import requests __UpperCAmelCase :Union[str, Any] = "https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=" def _a ( _lowercase : str ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page['''articles'''] , 1 ): print(F'{i}.) {article["title"]}' ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key="<Your BBC News API key goes here>")
240
0
'''simple docstring''' def _A (lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> int: '''simple docstring''' return number | (1 << position) def _A (lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> int: '''simple docstring''' return number & ~(1 << position) def _A (lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> int: '''simple docstring''' return number ^ (1 << position) def _A (lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> bool: '''simple docstring''' return ((number >> position) & 1) == 1 def _A (lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> int: '''simple docstring''' return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
168
'''simple docstring''' from __future__ import annotations from collections.abc import Generator def _A () -> Generator[int, None, None]: '''simple docstring''' _a = {} _a = 2 while True: _a = factor_map.pop(lowerCAmelCase__ , lowerCAmelCase__ ) if factor: _a = factor + prime while x in factor_map: x += factor _a = factor else: _a = prime yield prime prime += 1 def _A (lowerCAmelCase__ :float = 1E10 ) -> int: '''simple docstring''' _a = sieve() _a = 1 while True: _a = next(lowerCAmelCase__ ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(lowerCAmelCase__ ) n += 2 if __name__ == "__main__": print(solution())
168
1
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging __UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase ( lowerCAmelCase__ ): def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> Dict: super().__init__() if hasattr(scheduler.config, 'steps_offset') and scheduler.config.steps_offset != 1: snake_case_ = ( f'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' f' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate('steps_offset!=1', '1.0.0', lowerCAmelCase__, standard_warn=lowerCAmelCase__) snake_case_ = dict(scheduler.config) snake_case_ = 1 snake_case_ = FrozenDict(lowerCAmelCase__) if hasattr(scheduler.config, 'skip_prk_steps') and scheduler.config.skip_prk_steps is False: snake_case_ = ( f'The configuration file of this scheduler: {scheduler} has not set the configuration' ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate('skip_prk_steps not set', '1.0.0', lowerCAmelCase__, standard_warn=lowerCAmelCase__) snake_case_ = dict(scheduler.config) snake_case_ = True snake_case_ = FrozenDict(lowerCAmelCase__) if safety_checker is None: logger.warning( f'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .') self.register_modules( segmentation_model=lowerCAmelCase__, segmentation_processor=lowerCAmelCase__, vae=lowerCAmelCase__, text_encoder=lowerCAmelCase__, tokenizer=lowerCAmelCase__, unet=lowerCAmelCase__, scheduler=lowerCAmelCase__, safety_checker=lowerCAmelCase__, feature_extractor=lowerCAmelCase__, ) def a_ ( self, lowerCAmelCase__ = "auto") -> Optional[Any]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory snake_case_ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCAmelCase__) def a_ ( self) -> List[Any]: self.enable_attention_slicing(lowerCAmelCase__) def a_ ( self) -> Dict: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`') snake_case_ = torch.device('cuda') for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCAmelCase__, lowerCAmelCase__) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a_ ( self) -> Union[str, Any]: if self.device != torch.device('meta') or not hasattr(self.unet, '_hf_hook'): return self.device for module in self.unet.modules(): if ( hasattr(lowerCAmelCase__, '_hf_hook') and hasattr(module._hf_hook, 'execution_device') and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device) return self.device @torch.no_grad() def __call__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__ = 512, lowerCAmelCase__ = 512, lowerCAmelCase__ = 50, lowerCAmelCase__ = 7.5, lowerCAmelCase__ = None, lowerCAmelCase__ = 1, lowerCAmelCase__ = 0.0, lowerCAmelCase__ = None, lowerCAmelCase__ = None, lowerCAmelCase__ = "pil", lowerCAmelCase__ = True, lowerCAmelCase__ = None, lowerCAmelCase__ = 1, **lowerCAmelCase__, ) -> str: snake_case_ = self.segmentation_processor( text=[text], images=[image], padding='max_length', return_tensors='pt').to(self.device) snake_case_ = self.segmentation_model(**lowerCAmelCase__) snake_case_ = torch.sigmoid(outputs.logits).cpu().detach().unsqueeze(-1).numpy() snake_case_ = self.numpy_to_pil(lowerCAmelCase__)[0].resize(image.size) # Run inpainting pipeline with the generated mask snake_case_ = StableDiffusionInpaintPipeline( vae=self.vae, text_encoder=self.text_encoder, tokenizer=self.tokenizer, unet=self.unet, scheduler=self.scheduler, safety_checker=self.safety_checker, feature_extractor=self.feature_extractor, ) return inpainting_pipeline( prompt=lowerCAmelCase__, image=lowerCAmelCase__, mask_image=lowerCAmelCase__, height=lowerCAmelCase__, width=lowerCAmelCase__, num_inference_steps=lowerCAmelCase__, guidance_scale=lowerCAmelCase__, negative_prompt=lowerCAmelCase__, num_images_per_prompt=lowerCAmelCase__, eta=lowerCAmelCase__, generator=lowerCAmelCase__, latents=lowerCAmelCase__, output_type=lowerCAmelCase__, return_dict=lowerCAmelCase__, callback=lowerCAmelCase__, callback_steps=lowerCAmelCase__, )
312
"""simple docstring""" import os import numpy import onnx def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> List[str]: snake_case_ = a.name snake_case_ = b.name snake_case_ = '' snake_case_ = '' snake_case_ = a == b snake_case_ = name_a snake_case_ = name_b return res def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(UpperCAmelCase , UpperCAmelCase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , UpperCAmelCase , UpperCAmelCase ) _graph_replace_input_with(node_proto.attribute[1].g , UpperCAmelCase , UpperCAmelCase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , UpperCAmelCase , UpperCAmelCase ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for n in graph_proto.node: _node_replace_input_with(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Any: snake_case_ = list(model.graph.initializer ) snake_case_ = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i snake_case_ = inits[i].name snake_case_ = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , UpperCAmelCase , UpperCAmelCase ) def UpperCAmelCase ( UpperCAmelCase ) -> Optional[Any]: snake_case_ = os.path.dirname(UpperCAmelCase ) snake_case_ = os.path.basename(UpperCAmelCase ) snake_case_ = onnx.load(os.path.join(UpperCAmelCase , UpperCAmelCase ) ) snake_case_ = list(model.graph.initializer ) snake_case_ = set() snake_case_ = {} snake_case_ = [] snake_case_ = 0 for i in range(len(UpperCAmelCase ) ): if i in dup_set: continue for j in range(i + 1 , len(UpperCAmelCase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(UpperCAmelCase ) dup_set.add(UpperCAmelCase ) snake_case_ = inits[j].data_type snake_case_ = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , UpperCAmelCase ) total_reduced_size += mem_size snake_case_ = inits[i].name snake_case_ = inits[j].name if name_i in dup_map: dup_map[name_i].append(UpperCAmelCase ) else: snake_case_ = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) snake_case_ = sorted(UpperCAmelCase ) _remove_dup_initializers_from_model(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) snake_case_ = 'optimized_' + model_file_name snake_case_ = os.path.join(UpperCAmelCase , UpperCAmelCase ) onnx.save(UpperCAmelCase , UpperCAmelCase ) return new_model
312
1
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: lowercase__: Union[str, Any] = AutoConfig.from_pretrained(__UpperCAmelCase ) lowercase__: Any = FlaxAutoModelForSeqaSeqLM.from_config(config=__UpperCAmelCase ) lowercase__: List[str] = checkpoints.load_tax_checkpoint(__UpperCAmelCase ) lowercase__: Any = '''wi_0''' in tax_model['''target''']['''encoder''']['''layers_0''']['''mlp'''] if config.model_type == "t5": lowercase__: Any = '''SelfAttention''' if config.model_type == "longt5" and config.encoder_attention_type == "local": lowercase__: Tuple = '''LocalSelfAttention''' elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowercase__: Any = '''TransientGlobalSelfAttention''' else: raise ValueError( '''Given config is expected to have `model_type=\'t5\'`, or `model_type=\'longt5` with `encoder_attention_type`''' ''' attribute with a value from [\'local\', \'transient-global].''' ) # Encoder for layer_index in range(config.num_layers ): lowercase__: str = F"""layers_{str(__UpperCAmelCase )}""" # Self-Attention lowercase__: List[Any] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''key''']['''kernel'''] lowercase__: List[Any] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''out''']['''kernel'''] lowercase__: Optional[int] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''query''']['''kernel'''] lowercase__: Dict = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''value''']['''kernel'''] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowercase__: Any = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''T5LayerNorm_0''']['''scale'''] # Layer Normalization lowercase__: Union[str, Any] = tax_model['''target''']['''encoder'''][layer_name]['''pre_attention_layer_norm''']['''scale'''] if split_mlp_wi: lowercase__: List[str] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] lowercase__: List[Any] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: lowercase__: List[str] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] lowercase__: str = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization lowercase__: Any = tax_model['''target''']['''encoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning lowercase__: Optional[int] = flax_model.params['''encoder''']['''block'''][str(__UpperCAmelCase )]['''layer'''] lowercase__: Dict = tax_attention_key lowercase__: List[Any] = tax_attention_out lowercase__: Optional[int] = tax_attention_query lowercase__: Optional[Any] = tax_attention_value lowercase__: List[Any] = tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowercase__: int = tax_global_layer_norm if split_mlp_wi: lowercase__: int = tax_mlp_wi_a lowercase__: Optional[int] = tax_mlp_wi_a else: lowercase__: int = tax_mlp_wi lowercase__: int = tax_mlp_wo lowercase__: Tuple = tax_mlp_layer_norm lowercase__: int = flax_model_encoder_layer_block # Only for layer 0: lowercase__: List[str] = tax_model['''target''']['''encoder''']['''relpos_bias''']['''rel_embedding'''].T lowercase__: Optional[Any] = tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowercase__: int = tax_model['''target''']['''encoder''']['''side_relpos_bias''']['''rel_embedding'''].T lowercase__: Any = tax_encoder_global_rel_embedding # Assigning lowercase__: List[Any] = tax_model['''target''']['''encoder''']['''encoder_norm''']['''scale'''] lowercase__: Optional[Any] = tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): lowercase__: Tuple = F"""layers_{str(__UpperCAmelCase )}""" # Self-Attention lowercase__: Any = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''key''']['''kernel'''] lowercase__: Tuple = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''out''']['''kernel'''] lowercase__: Any = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''query''']['''kernel'''] lowercase__: Dict = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''value''']['''kernel'''] # Layer Normalization lowercase__: Dict = tax_model['''target''']['''decoder'''][layer_name]['''pre_self_attention_layer_norm'''][ '''scale''' ] # Encoder-Decoder-Attention lowercase__: str = tax_model['''target''']['''decoder'''][layer_name]['''encoder_decoder_attention'''] lowercase__: Any = tax_enc_dec_attention_module['''key''']['''kernel'''] lowercase__: Optional[Any] = tax_enc_dec_attention_module['''out''']['''kernel'''] lowercase__: int = tax_enc_dec_attention_module['''query''']['''kernel'''] lowercase__: Any = tax_enc_dec_attention_module['''value''']['''kernel'''] # Layer Normalization lowercase__: Any = tax_model['''target''']['''decoder'''][layer_name]['''pre_cross_attention_layer_norm''']['''scale'''] # MLP if split_mlp_wi: lowercase__: int = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] lowercase__: Any = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: lowercase__: Tuple = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] lowercase__: List[str] = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization lowercase__: Any = tax_model['''target''']['''decoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning lowercase__: Tuple = flax_model.params['''decoder''']['''block'''][str(__UpperCAmelCase )]['''layer'''] lowercase__: Tuple = tax_attention_key lowercase__: Any = tax_attention_out lowercase__: Dict = tax_attention_query lowercase__: Dict = tax_attention_value lowercase__: Any = tax_pre_attention_layer_norm lowercase__: Union[str, Any] = tax_enc_dec_attention_key lowercase__: Optional[int] = tax_enc_dec_attention_out lowercase__: Tuple = tax_enc_dec_attention_query lowercase__: Optional[Any] = tax_enc_dec_attention_value lowercase__: List[str] = tax_cross_layer_norm if split_mlp_wi: lowercase__: Optional[Any] = tax_mlp_wi_a lowercase__: Any = tax_mlp_wi_a else: lowercase__: List[Any] = tax_mlp_wi lowercase__: str = tax_mlp_wo lowercase__: List[Any] = txa_mlp_layer_norm lowercase__: str = flax_model_decoder_layer_block # Decoder Normalization lowercase__: Tuple = tax_model['''target''']['''decoder''']['''decoder_norm''']['''scale'''] lowercase__: Union[str, Any] = txa_decoder_norm # Only for layer 0: lowercase__: Tuple = tax_model['''target''']['''decoder''']['''relpos_bias''']['''rel_embedding'''].T lowercase__: Optional[int] = tax_decoder_rel_embedding # Token Embeddings lowercase__: List[str] = tax_model['''target''']['''token_embedder''']['''embedding'''] lowercase__: Union[str, Any] = txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: lowercase__: Dict = tax_model['''target''']['''decoder''']['''logits_dense''']['''kernel'''] flax_model.save_pretrained(__UpperCAmelCase ) print('''T5X Model was sucessfully converted!''' ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path the T5X checkpoint." ) parser.add_argument("--config_name", default=None, type=str, required=True, help="Config name of LongT5/T5 model.") parser.add_argument( "--flax_dump_folder_path", default=None, type=str, required=True, help="Path to the output FLAX model." ) __A = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
177
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class UpperCAmelCase (unittest.TestCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase=13 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=99 , _UpperCAmelCase=32 , _UpperCAmelCase=5 , _UpperCAmelCase=4 , _UpperCAmelCase=37 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=16 , _UpperCAmelCase=2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=4 , ): lowercase__: Dict = parent lowercase__: List[str] = batch_size lowercase__: Optional[Any] = seq_length lowercase__: List[Any] = is_training lowercase__: int = use_attention_mask lowercase__: Tuple = use_token_type_ids lowercase__: Union[str, Any] = use_labels lowercase__: str = vocab_size lowercase__: str = hidden_size lowercase__: str = num_hidden_layers lowercase__: Optional[int] = num_attention_heads lowercase__: List[str] = intermediate_size lowercase__: List[str] = hidden_act lowercase__: Tuple = hidden_dropout_prob lowercase__: int = attention_probs_dropout_prob lowercase__: int = max_position_embeddings lowercase__: Union[str, Any] = type_vocab_size lowercase__: List[Any] = type_sequence_label_size lowercase__: Any = initializer_range lowercase__: str = num_choices def _snake_case ( self ): lowercase__: int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__: List[Any] = None if self.use_attention_mask: lowercase__: Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__: List[Any] = None if self.use_token_type_ids: lowercase__: str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__: Optional[int] = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_UpperCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _snake_case ( self ): lowercase__: str = self.prepare_config_and_inputs() lowercase__, lowercase__, lowercase__, lowercase__: Optional[Any] = config_and_inputs lowercase__: Dict = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" _UpperCAmelCase :List[str] = True _UpperCAmelCase :Union[str, Any] = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def _snake_case ( self ): lowercase__: str = FlaxRoFormerModelTester(self ) @slow def _snake_case ( self ): for model_class_name in self.all_model_classes: lowercase__: Dict = model_class_name.from_pretrained('''junnyu/roformer_chinese_small''' , from_pt=_UpperCAmelCase ) lowercase__: int = model(np.ones((1, 1) ) ) self.assertIsNotNone(_UpperCAmelCase ) @require_flax class UpperCAmelCase (unittest.TestCase ): """simple docstring""" @slow def _snake_case ( self ): lowercase__: Any = FlaxRoFormerForMaskedLM.from_pretrained('''junnyu/roformer_chinese_base''' ) lowercase__: Optional[int] = jnp.array([[0, 1, 2, 3, 4, 5]] ) lowercase__: List[Any] = model(_UpperCAmelCase )[0] lowercase__: str = 50000 lowercase__: Tuple = (1, 6, vocab_size) self.assertEqual(output.shape , _UpperCAmelCase ) lowercase__: List[Any] = jnp.array( [[[-0.1_205, -1.0_265, 0.2_922], [-1.5_134, 0.1_974, 0.1_519], [-5.0_135, -3.9_003, -0.8_404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 ) )
177
1
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml __snake_case : Optional[Any] = NewType('DataClass', Any) __snake_case : int = NewType('DataClassType', Any) def _lowercase ( __snake_case ) -> List[Any]: if isinstance(__snake_case ,__snake_case ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def _lowercase ( __snake_case ) -> Callable[[str], Any]: __lowerCAmelCase : Any = {str(__snake_case ): choice for choice in choices} return lambda __snake_case : str_to_choice.get(__snake_case ,__snake_case ) def _lowercase ( *, __snake_case = None ,__snake_case = None ,__snake_case = dataclasses.MISSING ,__snake_case = dataclasses.MISSING ,__snake_case = None ,**__snake_case ,) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls __lowerCAmelCase : List[Any] = {} if aliases is not None: __lowerCAmelCase : Any = aliases if help is not None: __lowerCAmelCase : List[str] = help return dataclasses.field(metadata=__snake_case ,default=__snake_case ,default_factory=__snake_case ,**__snake_case ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 4_2 def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[DataClassType, Iterable[DataClassType]] , **_SCREAMING_SNAKE_CASE: Tuple) -> Any: """simple docstring""" if "formatter_class" not in kwargs: __lowerCAmelCase : Dict = ArgumentDefaultsHelpFormatter super().__init__(**_SCREAMING_SNAKE_CASE) if dataclasses.is_dataclass(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Union[str, Any] = [dataclass_types] __lowerCAmelCase : List[Any] = list(_SCREAMING_SNAKE_CASE) for dtype in self.dataclass_types: self._add_dataclass_arguments(_SCREAMING_SNAKE_CASE) @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE: ArgumentParser , _SCREAMING_SNAKE_CASE: dataclasses.Field) -> Optional[int]: """simple docstring""" __lowerCAmelCase : str = F"""--{field.name}""" __lowerCAmelCase : str = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , _SCREAMING_SNAKE_CASE): raise RuntimeError( "Unresolved type detected, which should have been done with the help of " "`typing.get_type_hints` method by default") __lowerCAmelCase : str = kwargs.pop("aliases" , []) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Any = [aliases] __lowerCAmelCase : Optional[int] = getattr(field.type , "__origin__" , field.type) if origin_type is Union or (hasattr(_SCREAMING_SNAKE_CASE , "UnionType") and isinstance(_SCREAMING_SNAKE_CASE , types.UnionType)): if str not in field.type.__args__ and ( len(field.type.__args__) != 2 or type(_SCREAMING_SNAKE_CASE) not in field.type.__args__ ): raise ValueError( "Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because" " the argument parser only supports one type per argument." F""" Problem encountered in field '{field.name}'.""") if type(_SCREAMING_SNAKE_CASE) not in field.type.__args__: # filter `str` in Union __lowerCAmelCase : List[str] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] __lowerCAmelCase : Any = getattr(field.type , "__origin__" , field.type) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) __lowerCAmelCase : Any = ( field.type.__args__[0] if isinstance(_SCREAMING_SNAKE_CASE , field.type.__args__[1]) else field.type.__args__[1] ) __lowerCAmelCase : Optional[int] = getattr(field.type , "__origin__" , field.type) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) __lowerCAmelCase : Tuple = {} if origin_type is Literal or (isinstance(field.type , _SCREAMING_SNAKE_CASE) and issubclass(field.type , _SCREAMING_SNAKE_CASE)): if origin_type is Literal: __lowerCAmelCase : str = field.type.__args__ else: __lowerCAmelCase : Tuple = [x.value for x in field.type] __lowerCAmelCase : Union[str, Any] = make_choice_type_function(kwargs["choices"]) if field.default is not dataclasses.MISSING: __lowerCAmelCase : int = field.default else: __lowerCAmelCase : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument __lowerCAmelCase : Optional[int] = copy(_SCREAMING_SNAKE_CASE) # Hack because type=bool in argparse does not behave as we want. __lowerCAmelCase : int = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. __lowerCAmelCase : Optional[int] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way __lowerCAmelCase : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name __lowerCAmelCase : List[str] = "?" # This is the value that will get picked if we do --field_name (without value) __lowerCAmelCase : Tuple = True elif isclass(_SCREAMING_SNAKE_CASE) and issubclass(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Union[str, Any] = field.type.__args__[0] __lowerCAmelCase : Any = "+" if field.default_factory is not dataclasses.MISSING: __lowerCAmelCase : str = field.default_factory() elif field.default is dataclasses.MISSING: __lowerCAmelCase : Optional[Any] = True else: __lowerCAmelCase : List[str] = field.type if field.default is not dataclasses.MISSING: __lowerCAmelCase : Optional[int] = field.default elif field.default_factory is not dataclasses.MISSING: __lowerCAmelCase : Union[str, Any] = field.default_factory() else: __lowerCAmelCase : Union[str, Any] = True parser.add_argument(_SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): __lowerCAmelCase : int = False parser.add_argument(F"""--no_{field.name}""" , action="store_false" , dest=field.name , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: DataClassType) -> List[Any]: """simple docstring""" if hasattr(_SCREAMING_SNAKE_CASE , "_argument_group_name"): __lowerCAmelCase : List[Any] = self.add_argument_group(dtype._argument_group_name) else: __lowerCAmelCase : Union[str, Any] = self try: __lowerCAmelCase : Dict[str, type] = get_type_hints(_SCREAMING_SNAKE_CASE) except NameError: raise RuntimeError( F"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ "removing line of `from __future__ import annotations` which opts in Postponed " "Evaluation of Annotations (PEP 563)") except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Optional[Any] = ".".join(map(_SCREAMING_SNAKE_CASE , sys.version_info[:3])) raise RuntimeError( F"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ "line of `from __future__ import annotations` which opts in union types as " "`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To " "support Python versions that lower than 3.10, you need to use " "`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of " "`X | None`.") from ex raise for field in dataclasses.fields(_SCREAMING_SNAKE_CASE): if not field.init: continue __lowerCAmelCase : List[Any] = type_hints[field.name] self._parse_dataclass_field(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: List[Any]=False , _SCREAMING_SNAKE_CASE: Any=True , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: str=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv)): __lowerCAmelCase : str = [] if args_filename: args_files.append(Path(_SCREAMING_SNAKE_CASE)) elif look_for_args_file and len(sys.argv): args_files.append(Path(sys.argv[0]).with_suffix(".args")) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values __lowerCAmelCase : Optional[int] = ArgumentParser() args_file_parser.add_argument(_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , action="append") # Use only remaining args for further parsing (remove the args_file_flag) __lowerCAmelCase : Optional[int] = args_file_parser.parse_known_args(args=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vars(_SCREAMING_SNAKE_CASE).get(args_file_flag.lstrip("-") , _SCREAMING_SNAKE_CASE) if cmd_args_file_paths: args_files.extend([Path(_SCREAMING_SNAKE_CASE) for p in cmd_args_file_paths]) __lowerCAmelCase : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last __lowerCAmelCase : int = file_args + args if args is not None else file_args + sys.argv[1:] __lowerCAmelCase : Union[str, Any] = self.parse_known_args(args=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = [] for dtype in self.dataclass_types: __lowerCAmelCase : int = {f.name for f in dataclasses.fields(_SCREAMING_SNAKE_CASE) if f.init} __lowerCAmelCase : Tuple = {k: v for k, v in vars(_SCREAMING_SNAKE_CASE).items() if k in keys} for k in keys: delattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = dtype(**_SCREAMING_SNAKE_CASE) outputs.append(_SCREAMING_SNAKE_CASE) if len(namespace.__dict__) > 0: # additional namespace. outputs.append(_SCREAMING_SNAKE_CASE) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""") return (*outputs,) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Dict[str, Any] , _SCREAMING_SNAKE_CASE: bool = False) -> Tuple[DataClass, ...]: """simple docstring""" __lowerCAmelCase : str = set(args.keys()) __lowerCAmelCase : str = [] for dtype in self.dataclass_types: __lowerCAmelCase : Optional[int] = {f.name for f in dataclasses.fields(_SCREAMING_SNAKE_CASE) if f.init} __lowerCAmelCase : Dict = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys()) __lowerCAmelCase : List[str] = dtype(**_SCREAMING_SNAKE_CASE) outputs.append(_SCREAMING_SNAKE_CASE) if not allow_extra_keys and unused_keys: raise ValueError(F"""Some keys are not used by the HfArgumentParser: {sorted(_SCREAMING_SNAKE_CASE)}""") return tuple(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: bool = False) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(_SCREAMING_SNAKE_CASE) , encoding="utf-8") as open_json_file: __lowerCAmelCase : Any = json.loads(open_json_file.read()) __lowerCAmelCase : str = self.parse_dict(_SCREAMING_SNAKE_CASE , allow_extra_keys=_SCREAMING_SNAKE_CASE) return tuple(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: bool = False) -> Tuple[DataClass, ...]: """simple docstring""" __lowerCAmelCase : Tuple = self.parse_dict(yaml.safe_load(Path(_SCREAMING_SNAKE_CASE).read_text()) , allow_extra_keys=_SCREAMING_SNAKE_CASE) return tuple(_SCREAMING_SNAKE_CASE)
351
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
58
0
import functools def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = len(A__ ) __lowerCamelCase = len(A__ ) @functools.cache def min_distance(A__ : int , A__ : int ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa __lowerCamelCase = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , A__ ) , 1 + min_distance(A__ , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
12
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class snake_case ( unittest.TestCase ): def lowercase_ ( self : Optional[int])-> int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowercase_ ( self : int)-> str: '''simple docstring''' __lowerCAmelCase: str = 1 __lowerCAmelCase: Union[str, Any] = 3 __lowerCAmelCase: Union[str, Any] = (3_2, 3_2) __lowerCAmelCase: Union[str, Any] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(UpperCamelCase__) return image @property def lowercase_ ( self : Tuple)-> str: '''simple docstring''' torch.manual_seed(0) __lowerCAmelCase: Optional[Any] = UNetaDConditionModel( block_out_channels=(3_2, 3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=7 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=3_2 , attention_head_dim=8 , use_linear_projection=UpperCamelCase__ , only_cross_attention=(True, True, False) , num_class_embeds=1_0_0 , ) return model @property def lowercase_ ( self : Any)-> Optional[Any]: '''simple docstring''' torch.manual_seed(0) __lowerCAmelCase: Tuple = AutoencoderKL( block_out_channels=[3_2, 3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def lowercase_ ( self : Any)-> Optional[Any]: '''simple docstring''' torch.manual_seed(0) __lowerCAmelCase: List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="gelu" , projection_dim=5_1_2 , ) return CLIPTextModel(UpperCamelCase__) def lowercase_ ( self : List[str])-> Dict: '''simple docstring''' __lowerCAmelCase: Tuple = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase: int = self.dummy_cond_unet_upscale __lowerCAmelCase: int = DDPMScheduler() __lowerCAmelCase: List[str] = DDIMScheduler(prediction_type="v_prediction") __lowerCAmelCase: Tuple = self.dummy_vae __lowerCAmelCase: Optional[Any] = self.dummy_text_encoder __lowerCAmelCase: Any = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") __lowerCAmelCase: Tuple = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] __lowerCAmelCase: List[Any] = Image.fromarray(np.uinta(UpperCamelCase__)).convert("RGB").resize((6_4, 6_4)) # make sure here that pndm scheduler skips prk __lowerCAmelCase: Optional[int] = StableDiffusionUpscalePipeline( unet=UpperCamelCase__ , low_res_scheduler=UpperCamelCase__ , scheduler=UpperCamelCase__ , vae=UpperCamelCase__ , text_encoder=UpperCamelCase__ , tokenizer=UpperCamelCase__ , max_noise_level=3_5_0 , ) __lowerCAmelCase: Tuple = sd_pipe.to(UpperCamelCase__) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__) __lowerCAmelCase: Any = "A painting of a squirrel eating a burger" __lowerCAmelCase: str = torch.Generator(device=UpperCamelCase__).manual_seed(0) __lowerCAmelCase: Optional[int] = sd_pipe( [prompt] , image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type="np" , ) __lowerCAmelCase: List[str] = output.images __lowerCAmelCase: Union[str, Any] = torch.Generator(device=UpperCamelCase__).manual_seed(0) __lowerCAmelCase: List[str] = sd_pipe( [prompt] , image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type="np" , return_dict=UpperCamelCase__ , )[0] __lowerCAmelCase: int = image[0, -3:, -3:, -1] __lowerCAmelCase: Dict = image_from_tuple[0, -3:, -3:, -1] __lowerCAmelCase: Dict = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) __lowerCAmelCase: List[Any] = np.array([0.3113, 0.3910, 0.4272, 0.4859, 0.5061, 0.4652, 0.5362, 0.5715, 0.5661]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def lowercase_ ( self : List[str])-> Optional[Any]: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase: Dict = self.dummy_cond_unet_upscale __lowerCAmelCase: List[str] = DDPMScheduler() __lowerCAmelCase: Union[str, Any] = DDIMScheduler(prediction_type="v_prediction") __lowerCAmelCase: Optional[int] = self.dummy_vae __lowerCAmelCase: List[Any] = self.dummy_text_encoder __lowerCAmelCase: Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") __lowerCAmelCase: List[Any] = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] __lowerCAmelCase: str = Image.fromarray(np.uinta(UpperCamelCase__)).convert("RGB").resize((6_4, 6_4)) # make sure here that pndm scheduler skips prk __lowerCAmelCase: Optional[int] = StableDiffusionUpscalePipeline( unet=UpperCamelCase__ , low_res_scheduler=UpperCamelCase__ , scheduler=UpperCamelCase__ , vae=UpperCamelCase__ , text_encoder=UpperCamelCase__ , tokenizer=UpperCamelCase__ , max_noise_level=3_5_0 , ) __lowerCAmelCase: Optional[int] = sd_pipe.to(UpperCamelCase__) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__) __lowerCAmelCase: List[str] = "A painting of a squirrel eating a burger" __lowerCAmelCase: List[Any] = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type="np" , ) __lowerCAmelCase: List[Any] = output.images assert image.shape[0] == 2 __lowerCAmelCase: Dict = torch.Generator(device=UpperCamelCase__).manual_seed(0) __lowerCAmelCase: Optional[Any] = sd_pipe( [prompt] , image=UpperCamelCase__ , generator=UpperCamelCase__ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type="np" , ) __lowerCAmelCase: List[Any] = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU") def lowercase_ ( self : Tuple)-> Any: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = self.dummy_cond_unet_upscale __lowerCAmelCase: int = DDPMScheduler() __lowerCAmelCase: int = DDIMScheduler(prediction_type="v_prediction") __lowerCAmelCase: Dict = self.dummy_vae __lowerCAmelCase: int = self.dummy_text_encoder __lowerCAmelCase: List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") __lowerCAmelCase: List[Any] = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] __lowerCAmelCase: Optional[int] = Image.fromarray(np.uinta(UpperCamelCase__)).convert("RGB").resize((6_4, 6_4)) # put models in fp16, except vae as it overflows in fp16 __lowerCAmelCase: List[Any] = unet.half() __lowerCAmelCase: List[str] = text_encoder.half() # make sure here that pndm scheduler skips prk __lowerCAmelCase: List[Any] = StableDiffusionUpscalePipeline( unet=UpperCamelCase__ , low_res_scheduler=UpperCamelCase__ , scheduler=UpperCamelCase__ , vae=UpperCamelCase__ , text_encoder=UpperCamelCase__ , tokenizer=UpperCamelCase__ , max_noise_level=3_5_0 , ) __lowerCAmelCase: str = sd_pipe.to(UpperCamelCase__) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__) __lowerCAmelCase: Optional[Any] = "A painting of a squirrel eating a burger" __lowerCAmelCase: str = torch.manual_seed(0) __lowerCAmelCase: Dict = sd_pipe( [prompt] , image=UpperCamelCase__ , generator=UpperCamelCase__ , num_inference_steps=2 , output_type="np" , ).images __lowerCAmelCase: Optional[Any] = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class snake_case ( unittest.TestCase ): def lowercase_ ( self : Tuple)-> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self : List[Any])-> Tuple: '''simple docstring''' __lowerCAmelCase: Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") __lowerCAmelCase: Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat.npy") __lowerCAmelCase: str = "stabilityai/stable-diffusion-x4-upscaler" __lowerCAmelCase: Optional[int] = StableDiffusionUpscalePipeline.from_pretrained(UpperCamelCase__) pipe.to(UpperCamelCase__) pipe.set_progress_bar_config(disable=UpperCamelCase__) pipe.enable_attention_slicing() __lowerCAmelCase: Tuple = "a cat sitting on a park bench" __lowerCAmelCase: int = torch.manual_seed(0) __lowerCAmelCase: List[Any] = pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , generator=UpperCamelCase__ , output_type="np" , ) __lowerCAmelCase: Dict = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 1e-3 def lowercase_ ( self : Optional[int])-> Any: '''simple docstring''' __lowerCAmelCase: Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") __lowerCAmelCase: Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat_fp16.npy") __lowerCAmelCase: Optional[Any] = "stabilityai/stable-diffusion-x4-upscaler" __lowerCAmelCase: Tuple = StableDiffusionUpscalePipeline.from_pretrained( UpperCamelCase__ , torch_dtype=torch.floataa , ) pipe.to(UpperCamelCase__) pipe.set_progress_bar_config(disable=UpperCamelCase__) pipe.enable_attention_slicing() __lowerCAmelCase: str = "a cat sitting on a park bench" __lowerCAmelCase: List[str] = torch.manual_seed(0) __lowerCAmelCase: Optional[Any] = pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , generator=UpperCamelCase__ , output_type="np" , ) __lowerCAmelCase: Union[str, Any] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 5e-1 def lowercase_ ( self : Optional[int])-> Dict: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase: Tuple = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") __lowerCAmelCase: Union[str, Any] = "stabilityai/stable-diffusion-x4-upscaler" __lowerCAmelCase: Any = StableDiffusionUpscalePipeline.from_pretrained( UpperCamelCase__ , torch_dtype=torch.floataa , ) pipe.to(UpperCamelCase__) pipe.set_progress_bar_config(disable=UpperCamelCase__) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() __lowerCAmelCase: int = "a cat sitting on a park bench" __lowerCAmelCase: Dict = torch.manual_seed(0) __lowerCAmelCase: Dict = pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , generator=UpperCamelCase__ , num_inference_steps=5 , output_type="np" , ) __lowerCAmelCase: Optional[int] = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 1_0**9
217
0
import requests from bsa import BeautifulSoup def a__ ( __UpperCamelCase = "AAPL" ): SCREAMING_SNAKE_CASE_ = F'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' SCREAMING_SNAKE_CASE_ = BeautifulSoup(requests.get(__UpperCamelCase ).text , "html.parser" ) SCREAMING_SNAKE_CASE_ = "My(6px) Pos(r) smartphone_Mt(6px)" return soup.find("div" , class_=class_ ).find("span" ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f"Current {symbol:<4} stock price is {stock_price(symbol):>8}")
359
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable A : List[Any] = {"configuration_dpt": ["DPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DPTConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Optional[Any] = ["DPTFeatureExtractor"] A : str = ["DPTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Optional[Any] = [ "DPT_PRETRAINED_MODEL_ARCHIVE_LIST", "DPTForDepthEstimation", "DPTForSemanticSegmentation", "DPTModel", "DPTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys A : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
305
0
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys A__ = subprocess.check_output("""git merge-base main HEAD""".split()).decode("""utf-8""") A__ = subprocess.check_output(f"git diff --name-only {fork_point_sha}".split()).decode("""utf-8""").split() A__ = """|""".join(sys.argv[1:]) A__ = re.compile(Rf"^({joined_dirs}).*?\.py$") A__ = [x for x in modified_files if regex.match(x)] print(""" """.join(relevant_modified_files), end="""""")
82
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( __A , __A , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Union[str, Any] = StableDiffusionXLImgaImgPipeline __UpperCamelCase : str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} __UpperCamelCase : List[str] = PipelineTesterMixin.required_optional_params - {'latents'} __UpperCamelCase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __UpperCamelCase : Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS __UpperCamelCase : int = IMAGE_TO_IMAGE_IMAGE_PARAMS def _snake_case (self ): torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__lowercase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCAmelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCAmelCase = CLIPTextModel(__lowercase ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__lowercase ) __lowerCAmelCase = CLIPTextModelWithProjection(__lowercase ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__lowercase ) __lowerCAmelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def _snake_case (self , __lowercase , __lowercase=0 ): __lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase ) ).to(__lowercase ) __lowerCAmelCase = image / 2 + 0.5 if str(__lowercase ).startswith('''mps''' ): __lowerCAmelCase = torch.manual_seed(__lowercase ) else: __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = sd_pipe(**__lowercase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _snake_case (self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def _snake_case (self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def _snake_case (self ): pass def _snake_case (self ): __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) # forward without prompt embeds __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 3 * ['''this is a negative prompt'''] __lowerCAmelCase = negative_prompt __lowerCAmelCase = 3 * [inputs['''prompt''']] __lowerCAmelCase = sd_pipe(**__lowercase ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 3 * ['''this is a negative prompt'''] __lowerCAmelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = sd_pipe.encode_prompt(__lowercase , negative_prompt=__lowercase ) __lowerCAmelCase = sd_pipe( **__lowercase , prompt_embeds=__lowercase , negative_prompt_embeds=__lowercase , pooled_prompt_embeds=__lowercase , negative_pooled_prompt_embeds=__lowercase , ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case (self , __lowercase , __lowercase="cpu" , __lowercase=torch.floataa , __lowercase=0 ): __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = np.random.RandomState(__lowercase ).standard_normal((1, 4, 64, 64) ) __lowerCAmelCase = torch.from_numpy(__lowercase ).to(device=__lowercase , dtype=__lowercase ) __lowerCAmelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def _snake_case (self ): __lowerCAmelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs(__lowercase ) __lowerCAmelCase = pipe(**__lowercase ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCAmelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
174
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''shi-labs/nat-mini-in1k-224''': '''https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json''', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "nat" SCREAMING_SNAKE_CASE_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self, lowerCAmelCase__=4, lowerCAmelCase__=3, lowerCAmelCase__=64, lowerCAmelCase__=[3, 4, 6, 5], lowerCAmelCase__=[2, 4, 8, 16], lowerCAmelCase__=7, lowerCAmelCase__=3.0, lowerCAmelCase__=True, lowerCAmelCase__=0.0, lowerCAmelCase__=0.0, lowerCAmelCase__=0.1, lowerCAmelCase__="gelu", lowerCAmelCase__=0.02, lowerCAmelCase__=1e-5, lowerCAmelCase__=0.0, lowerCAmelCase__=None, lowerCAmelCase__=None, **lowerCAmelCase__, ) -> Optional[Any]: super().__init__(**lowerCAmelCase__) snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = len(lowerCAmelCase__) snake_case_ = num_heads snake_case_ = kernel_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = layer_norm_eps snake_case_ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model snake_case_ = int(embed_dim * 2 ** (len(lowerCAmelCase__) - 1)) snake_case_ = layer_scale_init_value snake_case_ = ['stem'] + [f'stage{idx}' for idx in range(1, len(lowerCAmelCase__) + 1)] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=lowerCAmelCase__, out_indices=lowerCAmelCase__, stage_names=self.stage_names)
352
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''CarlCochet/trajectory-transformer-halfcheetah-medium-v2''': ( '''https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json''' ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "trajectory_transformer" SCREAMING_SNAKE_CASE_ = ["past_key_values"] SCREAMING_SNAKE_CASE_ = { "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self, lowerCAmelCase__=100, lowerCAmelCase__=5, lowerCAmelCase__=1, lowerCAmelCase__=1, lowerCAmelCase__=249, lowerCAmelCase__=6, lowerCAmelCase__=17, lowerCAmelCase__=25, lowerCAmelCase__=4, lowerCAmelCase__=4, lowerCAmelCase__=128, lowerCAmelCase__=0.1, lowerCAmelCase__=0.1, lowerCAmelCase__=0.1, lowerCAmelCase__=0.0006, lowerCAmelCase__=512, lowerCAmelCase__=0.02, lowerCAmelCase__=1e-12, lowerCAmelCase__=1, lowerCAmelCase__=True, lowerCAmelCase__=1, lowerCAmelCase__=5_0256, lowerCAmelCase__=5_0256, **lowerCAmelCase__, ) -> Optional[Any]: snake_case_ = vocab_size snake_case_ = action_weight snake_case_ = reward_weight snake_case_ = value_weight snake_case_ = max_position_embeddings snake_case_ = block_size snake_case_ = action_dim snake_case_ = observation_dim snake_case_ = transition_dim snake_case_ = learning_rate snake_case_ = n_layer snake_case_ = n_head snake_case_ = n_embd snake_case_ = embd_pdrop snake_case_ = attn_pdrop snake_case_ = resid_pdrop snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = kaiming_initializer_range snake_case_ = use_cache super().__init__(pad_token_id=lowerCAmelCase__, bos_token_id=lowerCAmelCase__, eos_token_id=lowerCAmelCase__, **lowerCAmelCase__)
312
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __UpperCamelCase = { '''configuration_squeezebert''': [ '''SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SqueezeBertConfig''', '''SqueezeBertOnnxConfig''', ], '''tokenization_squeezebert''': ['''SqueezeBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''SqueezeBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SqueezeBertForMaskedLM''', '''SqueezeBertForMultipleChoice''', '''SqueezeBertForQuestionAnswering''', '''SqueezeBertForSequenceClassification''', '''SqueezeBertForTokenClassification''', '''SqueezeBertModel''', '''SqueezeBertModule''', '''SqueezeBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Union[str, Any] =RoCBertTokenizer lowercase_ : str =None lowercase_ : Optional[Any] =False lowercase_ : Any =True lowercase_ : int =filter_non_english def A__ ( self): super().setUp() lowercase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''你''', '''好''', '''是''', '''谁''', '''a''', '''b''', '''c''', '''d'''] lowercase = {} lowercase = {} for i, value in enumerate(A__): lowercase = i lowercase = i lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file''']) lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''word_shape_file''']) lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''word_pronunciation_file''']) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''') as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens])) with open(self.word_shape_file ,'''w''' ,encoding='''utf-8''') as word_shape_writer: json.dump(A__ ,A__ ,ensure_ascii=A__) with open(self.word_pronunciation_file ,'''w''' ,encoding='''utf-8''') as word_pronunciation_writer: json.dump(A__ ,A__ ,ensure_ascii=A__) def A__ ( self): lowercase = self.tokenizer_class(self.vocab_file ,self.word_shape_file ,self.word_pronunciation_file) lowercase = tokenizer.tokenize('''你好[SEP]你是谁''') self.assertListEqual(A__ ,['''你''', '''好''', '''[SEP]''', '''你''', '''是''', '''谁''']) self.assertListEqual(tokenizer.convert_tokens_to_ids(A__) ,[5, 6, 2, 5, 7, 8]) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(A__) ,[5, 6, 2, 5, 7, 8]) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(A__) ,[5, 6, 2, 5, 7, 8]) def A__ ( self): lowercase = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''') ,['''ah''', '''\u535A''', '''\u63A8''', '''zz''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''') ,['''hello''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') ,['''hello''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__ ,strip_accents=A__) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') ,['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') ,['''h\u00E9llo''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__ ,strip_accents=A__) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') ,['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') ,['''hello''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') ,['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') ,['''hello''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''') ,['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__ ,strip_accents=A__) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') ,['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__ ,strip_accents=A__) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') ,['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def A__ ( self): lowercase = RoCBertBasicTokenizer(do_lower_case=A__ ,never_split=['''[UNK]''']) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''') ,['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]''']) def A__ ( self): lowercase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] lowercase = {} for i, token in enumerate(A__): lowercase = i lowercase = RoCBertWordpieceTokenizer(vocab=A__ ,unk_token='''[UNK]''') self.assertListEqual(tokenizer.tokenize('''''') ,[]) self.assertListEqual(tokenizer.tokenize('''unwanted running''') ,['''un''', '''##want''', '''##ed''', '''runn''', '''##ing''']) self.assertListEqual(tokenizer.tokenize('''unwantedX running''') ,['''[UNK]''', '''runn''', '''##ing''']) def A__ ( self): self.assertTrue(_is_whitespace(''' ''')) self.assertTrue(_is_whitespace('''\t''')) self.assertTrue(_is_whitespace('''\r''')) self.assertTrue(_is_whitespace('''\n''')) self.assertTrue(_is_whitespace('''\u00A0''')) self.assertFalse(_is_whitespace('''A''')) self.assertFalse(_is_whitespace('''-''')) def A__ ( self): self.assertTrue(_is_control('''\u0005''')) self.assertFalse(_is_control('''A''')) self.assertFalse(_is_control(''' ''')) self.assertFalse(_is_control('''\t''')) self.assertFalse(_is_control('''\r''')) def A__ ( self): self.assertTrue(_is_punctuation('''-''')) self.assertTrue(_is_punctuation('''$''')) self.assertTrue(_is_punctuation('''`''')) self.assertTrue(_is_punctuation('''.''')) self.assertFalse(_is_punctuation('''A''')) self.assertFalse(_is_punctuation(''' ''')) def A__ ( self): lowercase = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A__) for t in ['''Test''', '''\xad''', '''test''']] ,[['''[UNK]'''], [], ['''[UNK]''']]) if self.test_rust_tokenizer: lowercase = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(A__) for t in ['''Test''', '''\xad''', '''test''']] ,[['''[UNK]'''], [], ['''[UNK]''']]) def A__ ( self): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): lowercase = self.rust_tokenizer_class.from_pretrained(A__ ,**A__) lowercase = f'A, naïve {tokenizer_r.mask_token} AllenNLP sentence.' lowercase = tokenizer_r.encode_plus( A__ ,return_attention_mask=A__ ,return_token_type_ids=A__ ,return_offsets_mapping=A__ ,add_special_tokens=A__ ,) lowercase = tokenizer_r.do_lower_case if hasattr(A__ ,'''do_lower_case''') else False lowercase = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''A'''), ((1, 2), ''','''), ((3, 5), '''na'''), ((5, 6), '''##ï'''), ((6, 8), '''##ve'''), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), '''Allen'''), ((2_1, 2_3), '''##NL'''), ((2_3, 2_4), '''##P'''), ((2_5, 3_3), '''sentence'''), ((3_3, 3_4), '''.'''), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''a'''), ((1, 2), ''','''), ((3, 8), '''naive'''), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), '''allen'''), ((2_1, 2_3), '''##nl'''), ((2_3, 2_4), '''##p'''), ((2_5, 3_3), '''sentence'''), ((3_3, 3_4), '''.'''), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] ,tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''])) self.assertEqual([e[0] for e in expected_results] ,tokens['''offset_mapping''']) def A__ ( self): lowercase = ['''的''', '''人''', '''有'''] lowercase = ''''''.join(A__) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): lowercase = True lowercase = self.tokenizer_class.from_pretrained(A__ ,**A__) lowercase = self.rust_tokenizer_class.from_pretrained(A__ ,**A__) lowercase = tokenizer_p.encode(A__ ,add_special_tokens=A__) lowercase = tokenizer_r.encode(A__ ,add_special_tokens=A__) lowercase = tokenizer_r.convert_ids_to_tokens(A__) lowercase = tokenizer_p.convert_ids_to_tokens(A__) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A__ ,A__) self.assertListEqual(A__ ,A__) lowercase = False lowercase = self.rust_tokenizer_class.from_pretrained(A__ ,**A__) lowercase = self.tokenizer_class.from_pretrained(A__ ,**A__) lowercase = tokenizer_r.encode(A__ ,add_special_tokens=A__) lowercase = tokenizer_p.encode(A__ ,add_special_tokens=A__) lowercase = tokenizer_r.convert_ids_to_tokens(A__) lowercase = tokenizer_p.convert_ids_to_tokens(A__) # it is expected that only the first Chinese character is not preceded by "##". lowercase = [ f'##{token}' if idx != 0 else token for idx, token in enumerate(A__) ] self.assertListEqual(A__ ,A__) self.assertListEqual(A__ ,A__) @slow def A__ ( self): lowercase = self.tokenizer_class(self.vocab_file ,self.word_shape_file ,self.word_pronunciation_file) lowercase = tokenizer.encode('''你好''' ,add_special_tokens=A__) lowercase = tokenizer.encode('''你是谁''' ,add_special_tokens=A__) lowercase = tokenizer.build_inputs_with_special_tokens(A__) lowercase = tokenizer.build_inputs_with_special_tokens(A__ ,A__) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def A__ ( self): lowercase = self.get_tokenizers(do_lower_case=A__) for tokenizer in tokenizers: with self.subTest(f'{tokenizer.__class__.__name__}'): lowercase = '''你好,你是谁''' lowercase = tokenizer.tokenize(A__) lowercase = tokenizer.convert_tokens_to_ids(A__) lowercase = tokenizer.convert_tokens_to_shape_ids(A__) lowercase = tokenizer.convert_tokens_to_pronunciation_ids(A__) lowercase = tokenizer.prepare_for_model( A__ ,A__ ,A__ ,add_special_tokens=A__) lowercase = tokenizer.encode_plus(A__ ,add_special_tokens=A__) self.assertEqual(A__ ,A__)
101
0
'''simple docstring''' def UpperCamelCase ( _lowerCamelCase : str ): A__ = [0 for i in range(len(_lowerCamelCase ) )] # initialize interval's left pointer and right pointer A__, A__ = 0, 0 for i in range(1 , len(_lowerCamelCase ) ): # case when current index is inside the interval if i <= right_pointer: A__ = min(right_pointer - i + 1 , z_result[i - left_pointer] ) A__ = min_edge while go_next(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: A__, A__ = i, i + z_result[i] - 1 return z_result def UpperCamelCase ( _lowerCamelCase : int , _lowerCamelCase : list[int] , _lowerCamelCase : str ): return i + z_result[i] < len(_lowerCamelCase ) and s[z_result[i]] == s[i + z_result[i]] def UpperCamelCase ( _lowerCamelCase : str , _lowerCamelCase : str ): A__ = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string A__ = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(_lowerCamelCase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
123
'''simple docstring''' from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class UpperCAmelCase : def __init__( self :str , lowercase_ :str , )-> str: A__ = parent A__ = 13 A__ = 7 A__ = True A__ = True A__ = False A__ = True A__ = 99 A__ = 32 A__ = 2 A__ = 4 A__ = 37 A__ = "gelu" A__ = 0.1 A__ = 0.1 A__ = 5_12 A__ = 16 A__ = 2 A__ = 0.0_2 A__ = 3 A__ = 4 A__ = None def UpperCAmelCase_ ( self :Union[str, Any] )-> int: A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ = ids_tensor([self.batch_size] , self.num_choices ) A__ = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self :str , lowercase_ :Optional[int] , lowercase_ :List[str] , lowercase_ :Any , lowercase_ :Union[str, Any] , lowercase_ :Optional[int] , lowercase_ :str )-> List[str]: A__ = TFDistilBertModel(config=lowercase_ ) A__ = {"input_ids": input_ids, "attention_mask": input_mask} A__ = model(lowercase_ ) A__ = [input_ids, input_mask] A__ = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self :List[str] , lowercase_ :str , lowercase_ :Optional[Any] , lowercase_ :Optional[int] , lowercase_ :Optional[int] , lowercase_ :Optional[int] , lowercase_ :Union[str, Any] )-> Optional[int]: A__ = TFDistilBertForMaskedLM(config=lowercase_ ) A__ = {"input_ids": input_ids, "attention_mask": input_mask} A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self :Any , lowercase_ :str , lowercase_ :str , lowercase_ :Optional[int] , lowercase_ :str , lowercase_ :List[Any] , lowercase_ :Union[str, Any] )-> Optional[int]: A__ = TFDistilBertForQuestionAnswering(config=lowercase_ ) A__ = { "input_ids": input_ids, "attention_mask": input_mask, } A__ = model(lowercase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self :Union[str, Any] , lowercase_ :Optional[int] , lowercase_ :Any , lowercase_ :Dict , lowercase_ :Tuple , lowercase_ :Optional[Any] , lowercase_ :Optional[int] )-> Any: A__ = self.num_labels A__ = TFDistilBertForSequenceClassification(lowercase_ ) A__ = {"input_ids": input_ids, "attention_mask": input_mask} A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self :str , lowercase_ :Optional[Any] , lowercase_ :List[Any] , lowercase_ :Dict , lowercase_ :Tuple , lowercase_ :int , lowercase_ :Union[str, Any] )-> str: A__ = self.num_choices A__ = TFDistilBertForMultipleChoice(lowercase_ ) A__ = tf.tile(tf.expand_dims(lowercase_ , 1 ) , (1, self.num_choices, 1) ) A__ = tf.tile(tf.expand_dims(lowercase_ , 1 ) , (1, self.num_choices, 1) ) A__ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self :str , lowercase_ :Any , lowercase_ :List[str] , lowercase_ :Any , lowercase_ :int , lowercase_ :List[Any] , lowercase_ :Tuple )-> Tuple: A__ = self.num_labels A__ = TFDistilBertForTokenClassification(lowercase_ ) A__ = {"input_ids": input_ids, "attention_mask": input_mask} A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self :Any )-> Union[str, Any]: A__ = self.prepare_config_and_inputs() ((A__), (A__), (A__), (A__), (A__), (A__)) = config_and_inputs A__ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): __lowercase = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) __lowercase = ( { """feature-extraction""": TFDistilBertModel, """fill-mask""": TFDistilBertForMaskedLM, """question-answering""": TFDistilBertForQuestionAnswering, """text-classification""": TFDistilBertForSequenceClassification, """token-classification""": TFDistilBertForTokenClassification, """zero-shot""": TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase = False __lowercase = False def UpperCAmelCase_ ( self :Optional[Any] )-> List[Any]: A__ = TFDistilBertModelTester(self ) A__ = ConfigTester(self , config_class=lowercase_ , dim=37 ) def UpperCAmelCase_ ( self :Tuple )-> Tuple: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self :int )-> Tuple: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*lowercase_ ) def UpperCAmelCase_ ( self :Optional[int] )-> Optional[Any]: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*lowercase_ ) def UpperCAmelCase_ ( self :str )-> str: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*lowercase_ ) def UpperCAmelCase_ ( self :List[str] )-> Dict: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*lowercase_ ) def UpperCAmelCase_ ( self :List[str] )-> Optional[int]: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*lowercase_ ) def UpperCAmelCase_ ( self :str )-> int: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*lowercase_ ) @slow def UpperCAmelCase_ ( self :List[str] )-> Dict: for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): A__ = TFDistilBertModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @require_tf class UpperCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase_ ( self :List[Any] )-> Any: A__ = TFDistilBertModel.from_pretrained("distilbert-base-uncased" ) A__ = tf.constant([[0, 1, 2, 3, 4, 5]] ) A__ = model(lowercase_ )[0] A__ = [1, 6, 7_68] self.assertEqual(output.shape , lowercase_ ) A__ = tf.constant( [ [ [0.1_9_2_6_1_8_8_5, -0.1_3_7_3_2_9_5_5, 0.4_1_1_9_7_9_9], [0.2_2_1_5_0_1_5_6, -0.0_7_4_2_2_6_6_1, 0.3_9_0_3_7_2_0_4], [0.2_2_7_5_6_0_1_8, -0.0_8_9_6_4_1_4, 0.3_7_0_1_4_6_7], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1E-4 )
123
1
def UpperCamelCase ( _A ): """simple docstring""" return sum(i for i in range(1, number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("Program to check whether a number is a Perfect number or not...") __magic_name__: Dict = int(input("Enter number: ").strip()) print(F"""{number} is {'' if perfect(number) else 'not '}a Perfect Number.""")
342
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class SCREAMING_SNAKE_CASE ( lowerCamelCase__ ): __lowerCamelCase : torch.FloatTensor class SCREAMING_SNAKE_CASE ( lowerCamelCase__ , lowerCamelCase__ ): @register_to_config def __init__( self : Dict , __lowercase : int = 32 , __lowercase : int = 64 , __lowercase : int = 20 , __lowercase : int = 768 , __lowercase : Any=77 , __lowercase : Optional[int]=4 , __lowercase : float = 0.0 , __lowercase : str = "silu" , __lowercase : Optional[str] = None , __lowercase : Optional[str] = None , __lowercase : Optional[str] = "linear" , __lowercase : Optional[str] = "prd" , __lowercase : Optional[int] = None , __lowercase : Optional[int] = None , __lowercase : Optional[int] = None , ): '''simple docstring''' super().__init__() __a = num_attention_heads __a = attention_head_dim __a = num_attention_heads * attention_head_dim __a = additional_embeddings __a = time_embed_dim or inner_dim __a = embedding_proj_dim or embedding_dim __a = clip_embed_dim or embedding_dim __a = Timesteps(__lowercase , __lowercase , 0 ) __a = TimestepEmbedding(__lowercase , __lowercase , out_dim=__lowercase , act_fn=__lowercase ) __a = nn.Linear(__lowercase , __lowercase ) if embedding_proj_norm_type is None: __a = None elif embedding_proj_norm_type == "layer": __a = nn.LayerNorm(__lowercase ) else: raise ValueError(F"unsupported embedding_proj_norm_type: {embedding_proj_norm_type}" ) __a = nn.Linear(__lowercase , __lowercase ) if encoder_hid_proj_type is None: __a = None elif encoder_hid_proj_type == "linear": __a = nn.Linear(__lowercase , __lowercase ) else: raise ValueError(F"unsupported encoder_hid_proj_type: {encoder_hid_proj_type}" ) __a = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , __lowercase ) ) if added_emb_type == "prd": __a = nn.Parameter(torch.zeros(1 , 1 , __lowercase ) ) elif added_emb_type is None: __a = None else: raise ValueError( F"`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`." ) __a = nn.ModuleList( [ BasicTransformerBlock( __lowercase , __lowercase , __lowercase , dropout=__lowercase , activation_fn="""gelu""" , attention_bias=__lowercase , ) for d in range(__lowercase ) ] ) if norm_in_type == "layer": __a = nn.LayerNorm(__lowercase ) elif norm_in_type is None: __a = None else: raise ValueError(F"Unsupported norm_in_type: {norm_in_type}." ) __a = nn.LayerNorm(__lowercase ) __a = nn.Linear(__lowercase , __lowercase ) __a = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0 ) causal_attention_mask.triu_(1 ) __a = causal_attention_mask[None, ...] self.register_buffer("""causal_attention_mask""" , __lowercase , persistent=__lowercase ) __a = nn.Parameter(torch.zeros(1 , __lowercase ) ) __a = nn.Parameter(torch.zeros(1 , __lowercase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' __a = {} def fn_recursive_add_processors(__lowercase : str , __lowercase : torch.nn.Module , __lowercase : Dict[str, AttentionProcessor] ): if hasattr(__lowercase , """set_processor""" ): __a = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"{name}.{sub_name}" , __lowercase , __lowercase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(__lowercase , __lowercase , __lowercase ) return processors def UpperCamelCase_ ( self : List[str] , __lowercase : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ): '''simple docstring''' __a = len(self.attn_processors.keys() ) if isinstance(__lowercase , __lowercase ) and len(__lowercase ) != count: raise ValueError( F"A dict of processors was passed, but the number of processors {len(__lowercase )} does not match the" F" number of attention layers: {count}. Please make sure to pass {count} processor classes." ) def fn_recursive_attn_processor(__lowercase : str , __lowercase : torch.nn.Module , __lowercase : Dict ): if hasattr(__lowercase , """set_processor""" ): if not isinstance(__lowercase , __lowercase ): module.set_processor(__lowercase ) else: module.set_processor(processor.pop(F"{name}.processor" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"{name}.{sub_name}" , __lowercase , __lowercase ) for name, module in self.named_children(): fn_recursive_attn_processor(__lowercase , __lowercase , __lowercase ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' self.set_attn_processor(AttnProcessor() ) def UpperCamelCase_ ( self : Union[str, Any] , __lowercase : Optional[int] , __lowercase : Union[torch.Tensor, float, int] , __lowercase : torch.FloatTensor , __lowercase : Optional[torch.FloatTensor] = None , __lowercase : Optional[torch.BoolTensor] = None , __lowercase : bool = True , ): '''simple docstring''' __a = hidden_states.shape[0] __a = timestep if not torch.is_tensor(__lowercase ): __a = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(__lowercase ) and len(timesteps.shape ) == 0: __a = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __a = timesteps * torch.ones(__lowercase , dtype=timesteps.dtype , device=timesteps.device ) __a = self.time_proj(__lowercase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. __a = timesteps_projected.to(dtype=self.dtype ) __a = self.time_embedding(__lowercase ) if self.embedding_proj_norm is not None: __a = self.embedding_proj_norm(__lowercase ) __a = self.embedding_proj(__lowercase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: __a = self.encoder_hidden_states_proj(__lowercase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("""`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set""" ) __a = self.proj_in(__lowercase ) __a = self.positional_embedding.to(hidden_states.dtype ) __a = [] __a = 0 if encoder_hidden_states is not None: additional_embeds.append(__lowercase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: __a = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: __a = hidden_states[:, None, :] __a = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: __a = self.prd_embedding.to(hidden_states.dtype ).expand(__lowercase , -1 , -1 ) additional_embeds.append(__lowercase ) __a = torch.cat( __lowercase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens __a = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: __a = F.pad( __lowercase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) __a = hidden_states + positional_embeddings if attention_mask is not None: __a = (1 - attention_mask.to(hidden_states.dtype )) * -10000.0 __a = F.pad(__lowercase , (0, self.additional_embeddings) , value=0.0 ) __a = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) __a = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: __a = self.norm_in(__lowercase ) for block in self.transformer_blocks: __a = block(__lowercase , attention_mask=__lowercase ) __a = self.norm_out(__lowercase ) if self.prd_embedding is not None: __a = hidden_states[:, -1] else: __a = hidden_states[:, additional_embeddings_len:] __a = self.proj_to_clip_embeddings(__lowercase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=__lowercase ) def UpperCamelCase_ ( self : Any , __lowercase : Tuple ): '''simple docstring''' __a = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
302
0
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) @add_end_docstrings(__a ) class lowerCAmelCase ( __a ): '''simple docstring''' def __init__( self : List[str] , *__a : Any , **__a : Any ) -> Any: """simple docstring""" super().__init__(*__lowerCAmelCase , **__lowerCAmelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def lowerCAmelCase ( self : Optional[int] , __a : Any=None ) -> Dict: """simple docstring""" __lowercase : Dict = {} if top_k is not None: __lowercase : Dict = top_k return {}, {}, postprocess_params def __call__( self : int , __a : Tuple , **__a : Any ) -> Tuple: """simple docstring""" return super().__call__(__lowerCAmelCase , **__lowerCAmelCase ) def lowerCAmelCase ( self : List[str] , __a : List[Any] ) -> List[str]: """simple docstring""" __lowercase : int = load_image(__lowerCAmelCase ) __lowercase : List[Any] = self.image_processor(images=__lowerCAmelCase , return_tensors=self.framework ) return model_inputs def lowerCAmelCase ( self : int , __a : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase : List[str] = self.model(**__lowerCAmelCase ) return model_outputs def lowerCAmelCase ( self : List[Any] , __a : Any , __a : Union[str, Any]=5 ) -> Any: """simple docstring""" if top_k > self.model.config.num_labels: __lowercase : Optional[int] = self.model.config.num_labels if self.framework == "pt": __lowercase : List[str] = model_outputs.logits.softmax(-1 )[0] __lowercase , __lowercase : Union[str, Any] = probs.topk(__lowerCAmelCase ) elif self.framework == "tf": __lowercase : int = stable_softmax(model_outputs.logits , axis=-1 )[0] __lowercase : str = tf.math.top_k(__lowerCAmelCase , k=__lowerCAmelCase ) __lowercase , __lowercase : int = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(F"Unsupported framework: {self.framework}" ) __lowercase : int = scores.tolist() __lowercase : List[str] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__lowerCAmelCase , __lowerCAmelCase )]
360
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class lowerCAmelCase : '''simple docstring''' def __init__( self : Optional[Any] , __a : Dict , __a : List[str]=None , __a : Optional[Any]=None , __a : Union[str, Any]=None , __a : int="resnet50" , __a : List[str]=3 , __a : Tuple=32 , __a : Dict=3 , __a : List[str]=True , __a : Union[str, Any]=True , ) -> Any: """simple docstring""" __lowercase : Optional[int] = parent __lowercase : List[str] = out_indices if out_indices is not None else [4] __lowercase : Optional[int] = stage_names __lowercase : Any = out_features __lowercase : Optional[Any] = backbone __lowercase : Optional[Any] = batch_size __lowercase : Union[str, Any] = image_size __lowercase : List[str] = num_channels __lowercase : str = use_pretrained_backbone __lowercase : str = is_training def lowerCAmelCase ( self : Dict ) -> Tuple: """simple docstring""" __lowercase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase : str = self.get_config() return config, pixel_values def lowerCAmelCase ( self : int ) -> str: """simple docstring""" return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def lowerCAmelCase ( self : Optional[int] , __a : Dict , __a : Any ) -> Dict: """simple docstring""" __lowercase : Dict = TimmBackbone(config=__a ) model.to(__a ) model.eval() with torch.no_grad(): __lowercase : Optional[Any] = model(__a ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def lowerCAmelCase ( self : Any ) -> int: """simple docstring""" __lowercase : Union[str, Any] = self.prepare_config_and_inputs() __lowercase , __lowercase : str = config_and_inputs __lowercase : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch @require_timm class lowerCAmelCase ( __a , __a , __a , unittest.TestCase ): '''simple docstring''' _A : List[Any] = (TimmBackbone,) if is_torch_available() else () _A : Dict = {'''feature-extraction''': TimmBackbone} if is_torch_available() else {} _A : List[Any] = False _A : List[str] = False _A : Any = False _A : Optional[Any] = False def lowerCAmelCase ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase : str = TimmBackboneModelTester(self ) __lowercase : Any = ConfigTester(self , config_class=__a , has_text_modality=__a ) def lowerCAmelCase ( self : Any ) -> str: """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase ( self : str ) -> Tuple: """simple docstring""" __lowercase : Tuple = """resnet18""" __lowercase : Optional[int] = """microsoft/resnet-18""" __lowercase : Union[str, Any] = AutoBackbone.from_pretrained(__a , use_timm_backbone=__a ) __lowercase : Dict = AutoBackbone.from_pretrained(__a ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) __lowercase : Union[str, Any] = AutoBackbone.from_pretrained(__a , use_timm_backbone=__a , out_indices=[1, 2, 3] ) __lowercase : Optional[Any] = AutoBackbone.from_pretrained(__a , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip("""TimmBackbone doesn't support feed forward chunking""" ) def lowerCAmelCase ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip("""TimmBackbone doesn't have num_hidden_layers attribute""" ) def lowerCAmelCase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip("""TimmBackbone initialization is managed on the timm side""" ) def lowerCAmelCase ( self : List[Any] ) -> str: """simple docstring""" pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def lowerCAmelCase ( self : Optional[int] ) -> Dict: """simple docstring""" pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def lowerCAmelCase ( self : Tuple ) -> Tuple: """simple docstring""" pass @unittest.skip("""TimmBackbone model cannot be created without specifying a backbone checkpoint""" ) def lowerCAmelCase ( self : Dict ) -> Optional[Any]: """simple docstring""" pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def lowerCAmelCase ( self : Union[str, Any] ) -> int: """simple docstring""" pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def lowerCAmelCase ( self : Union[str, Any] ) -> str: """simple docstring""" pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def lowerCAmelCase ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def lowerCAmelCase ( self : List[str] ) -> List[Any]: """simple docstring""" pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def lowerCAmelCase ( self : List[Any] ) -> Tuple: """simple docstring""" pass @unittest.skip("""TimmBackbone doesn't have hidden size info in its configuration.""" ) def lowerCAmelCase ( self : Dict ) -> Any: """simple docstring""" pass @unittest.skip("""TimmBackbone doesn't support output_attentions.""" ) def lowerCAmelCase ( self : str ) -> List[Any]: """simple docstring""" pass @unittest.skip("""Safetensors is not supported by timm.""" ) def lowerCAmelCase ( self : Any ) -> List[Any]: """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase ( self : List[str] ) -> List[str]: """simple docstring""" pass def lowerCAmelCase ( self : Any ) -> List[str]: """simple docstring""" __lowercase , __lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase : Optional[Any] = model_class(__a ) __lowercase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase : List[str] = [*signature.parameters.keys()] __lowercase : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __a ) def lowerCAmelCase ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase , __lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() __lowercase : Optional[Any] = True __lowercase : Union[str, Any] = self.has_attentions # no need to test all models as different heads yield the same functionality __lowercase : Union[str, Any] = self.all_model_classes[0] __lowercase : List[Any] = model_class(__a ) model.to(__a ) __lowercase : Optional[Any] = self._prepare_for_class(__a , __a ) __lowercase : Union[str, Any] = model(**__a ) __lowercase : Optional[int] = outputs[0][-1] # Encoder-/Decoder-only models __lowercase : Any = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: __lowercase : Optional[int] = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=__a ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def lowerCAmelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase , __lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase : List[str] = model_class(__a ) model.to(__a ) model.eval() __lowercase : int = model(**__a ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None __lowercase : Any = copy.deepcopy(__a ) __lowercase : Dict = None __lowercase : Tuple = model_class(__a ) model.to(__a ) model.eval() __lowercase : Optional[int] = model(**__a ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights __lowercase : List[str] = copy.deepcopy(__a ) __lowercase : Optional[Any] = False __lowercase : str = model_class(__a ) model.to(__a ) model.eval() __lowercase : List[Any] = model(**__a )
306
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __snake_case = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
176
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __snake_case = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class lowercase__ ( _UpperCAmelCase , unittest.TestCase ): A__ : str =DebertaVaTokenizer A__ : List[str] =DebertaVaTokenizerFast A__ : Any =True A__ : str =True def A_ ( self : Tuple ): super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , unk_token='<unk>' ) tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : Union[str, Any] , UpperCAmelCase_ : Optional[Any] ): SCREAMING_SNAKE_CASE__ = 'this is a test' SCREAMING_SNAKE_CASE__ = 'this is a test' return input_text, output_text def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = '<pad>' SCREAMING_SNAKE_CASE__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_ ) , UpperCAmelCase_ ) def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '[PAD]' ) self.assertEqual(len(UpperCAmelCase_ ) , 30001 ) def A_ ( self : str ): self.assertEqual(self.get_tokenizer().vocab_size , 30000 ) def A_ ( self : Optional[Any] ): # fmt: off SCREAMING_SNAKE_CASE__ = ' \tHeLLo!how \n Are yoU? ' SCREAMING_SNAKE_CASE__ = ['▁hello', '!', 'how', '▁are', '▁you', '?'] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.' ) def A_ ( self : Any ): pass @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.' ) def A_ ( self : Tuple ): pass def A_ ( self : List[str] ): # fmt: off SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : List[str] ): # fmt: off SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : int ): # fmt: off SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Tuple ): # fmt: off SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Any ): # fmt: off SCREAMING_SNAKE_CASE__ = ' \tHeLLo!how \n Are yoU? ' SCREAMING_SNAKE_CASE__ = ['▁', '<unk>', 'e', '<unk>', 'o', '!', 'how', '▁', '<unk>', 're', '▁yo', '<unk>', '?'] # fmt: on SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , split_by_punct=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : int ): SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = 'This is a test' SCREAMING_SNAKE_CASE__ = [13, 1, 4398, 25, 21, 1289] SCREAMING_SNAKE_CASE__ = ['▁', 'T', 'his', '▁is', '▁a', '▁test'] SCREAMING_SNAKE_CASE__ = ['▁', '<unk>', 'his', '▁is', '▁a', '▁test'] SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ , keep_accents=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast(UpperCAmelCase_ , keep_accents=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # fmt: off SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] SCREAMING_SNAKE_CASE__ = ['▁', 'I', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.', ] SCREAMING_SNAKE_CASE__ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('sequence builders' ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('multi-sequence build' ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , UpperCAmelCase_ ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , UpperCAmelCase_ , ) @slow def A_ ( self : Optional[Any] ): # fmt: off SCREAMING_SNAKE_CASE__ = {'input_ids': [[1, 39867, 36, 19390, 486, 27, 35052, 81436, 18, 60685, 1225, 7, 35052, 81436, 18, 9367, 16899, 18, 15937, 53, 594, 773, 18, 16287, 30465, 36, 15937, 6, 41139, 38, 36979, 60763, 191, 6, 34132, 99, 6, 50538, 390, 43230, 6, 34132, 2779, 20850, 14, 699, 1072, 1194, 36, 382, 10901, 53, 7, 699, 1072, 2084, 36, 20422, 630, 53, 19, 105, 3049, 1896, 1053, 16899, 1506, 11, 37978, 4243, 7, 1237, 31869, 200, 16566, 654, 6, 35052, 81436, 7, 55630, 13593, 4, 2], [1, 26, 15011, 13, 667, 8, 1053, 18, 23611, 1237, 72356, 12820, 34, 104134, 1209, 35, 13313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 15785, 14951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ , model_name='microsoft/deberta-v2-xlarge' , revision='ad6e42c1532ddf3a15c39246b63f5559d558b670' , )
176
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Any = { 'microsoft/swin-tiny-patch4-window7-224': ( 'https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json' ), # See all Swin models at https://huggingface.co/models?filter=swin } class __magic_name__ ( snake_case , snake_case ): UpperCamelCase_ :Any = """swin""" UpperCamelCase_ :int = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , _lowercase=224 , _lowercase=4 , _lowercase=3 , _lowercase=96 , _lowercase=[2, 2, 6, 2] , _lowercase=[3, 6, 12, 24] , _lowercase=7 , _lowercase=4.0 , _lowercase=True , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.1 , _lowercase="gelu" , _lowercase=False , _lowercase=0.02 , _lowercase=1e-5 , _lowercase=32 , _lowercase=None , _lowercase=None , **_lowercase , )-> Tuple: super().__init__(**_lowercase ) UpperCamelCase_ = image_size UpperCamelCase_ = patch_size UpperCamelCase_ = num_channels UpperCamelCase_ = embed_dim UpperCamelCase_ = depths UpperCamelCase_ = len(_lowercase ) UpperCamelCase_ = num_heads UpperCamelCase_ = window_size UpperCamelCase_ = mlp_ratio UpperCamelCase_ = qkv_bias UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = drop_path_rate UpperCamelCase_ = hidden_act UpperCamelCase_ = use_absolute_embeddings UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = initializer_range UpperCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCamelCase_ = int(embed_dim * 2 ** (len(_lowercase ) - 1) ) UpperCamelCase_ = ['''stem'''] + [F"stage{idx}" for idx in range(1 , len(_lowercase ) + 1 )] UpperCamelCase_ = get_aligned_output_features_output_indices( out_features=_lowercase , out_indices=_lowercase , stage_names=self.stage_names ) class __magic_name__ ( snake_case ): UpperCamelCase_ :Optional[int] = version.parse("""1.11""" ) @property def UpperCAmelCase_ ( self )-> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def UpperCAmelCase_ ( self )-> float: return 1e-4
351
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __magic_name__ : @staticmethod def UpperCAmelCase_ ( *_lowercase , **_lowercase )-> Optional[int]: pass @is_pipeline_test @require_vision class __magic_name__ ( unittest.TestCase ): @require_torch def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , ) UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) UpperCamelCase_ = image_classifier(_lowercase , candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(_lowercase ) , [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ] , ) UpperCamelCase_ = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], ] , ) @require_tf def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , framework="tf" ) UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) UpperCamelCase_ = image_classifier(_lowercase , candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(_lowercase ) , [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}] , ) UpperCamelCase_ = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], [ {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, {"score": 0.333, "label": ANY(_lowercase )}, ], ] , ) @slow @require_torch def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , ) # This is an image of 2 cats with remotes and no planes UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) UpperCamelCase_ = image_classifier(_lowercase , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(_lowercase ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) UpperCamelCase_ = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , ) @slow @require_tf def UpperCAmelCase_ ( self )-> List[str]: UpperCamelCase_ = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , framework="tf" ) # This is an image of 2 cats with remotes and no planes UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) UpperCamelCase_ = image_classifier(_lowercase , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(_lowercase ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) UpperCamelCase_ = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , )
60
0
'''simple docstring''' def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : set[int] = set() # To detect a back edge, keep track of vertices currently in the recursion stack UpperCAmelCase__ : set[int] = set() return any( node not in visited and depth_first_search(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) for node in graph ) def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): visited.add(UpperCamelCase__ ) rec_stk.add(UpperCamelCase__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(UpperCamelCase__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
163
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __A =logging.get_logger(__name__) __A ={ 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _snake_case ( a__ ): lowerCAmelCase :Optional[int] = '''pegasus''' lowerCAmelCase :Optional[int] = ['''past_key_values'''] lowerCAmelCase :str = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , _lowerCamelCase=5_0265 , _lowerCamelCase=1024 , _lowerCamelCase=12 , _lowerCamelCase=4096 , _lowerCamelCase=16 , _lowerCamelCase=12 , _lowerCamelCase=4096 , _lowerCamelCase=16 , _lowerCamelCase=0.0 , _lowerCamelCase=0.0 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="gelu" , _lowerCamelCase=1024 , _lowerCamelCase=0.1 , _lowerCamelCase=0.0 , _lowerCamelCase=0.0 , _lowerCamelCase=0.02 , _lowerCamelCase=0 , _lowerCamelCase=False , _lowerCamelCase=0 , _lowerCamelCase=1 , _lowerCamelCase=1 , **_lowerCamelCase , ): UpperCAmelCase__ : Union[str, Any] = vocab_size UpperCAmelCase__ : Union[str, Any] = max_position_embeddings UpperCAmelCase__ : List[Any] = d_model UpperCAmelCase__ : Union[str, Any] = encoder_ffn_dim UpperCAmelCase__ : Any = encoder_layers UpperCAmelCase__ : List[str] = encoder_attention_heads UpperCAmelCase__ : int = decoder_ffn_dim UpperCAmelCase__ : Any = decoder_layers UpperCAmelCase__ : Tuple = decoder_attention_heads UpperCAmelCase__ : Optional[int] = dropout UpperCAmelCase__ : Dict = attention_dropout UpperCAmelCase__ : Optional[int] = activation_dropout UpperCAmelCase__ : Dict = activation_function UpperCAmelCase__ : Optional[Any] = init_std UpperCAmelCase__ : int = encoder_layerdrop UpperCAmelCase__ : Tuple = decoder_layerdrop UpperCAmelCase__ : str = use_cache UpperCAmelCase__ : Any = encoder_layers UpperCAmelCase__ : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , is_encoder_decoder=_lowerCamelCase , decoder_start_token_id=_lowerCamelCase , forced_eos_token_id=_lowerCamelCase , **_lowerCamelCase , ) @property def snake_case__ ( self): return self.encoder_attention_heads @property def snake_case__ ( self): return self.d_model
163
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase : Optional[Any] = logging.get_logger(__name__) lowerCamelCase : List[Any] = { "facebook/deit-base-distilled-patch16-224": ( "https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json" ), # See all DeiT models at https://huggingface.co/models?filter=deit } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''deit''' def __init__( self : Optional[int] , A_ : str=768 , A_ : List[Any]=12 , A_ : Union[str, Any]=12 , A_ : Union[str, Any]=3072 , A_ : Any="gelu" , A_ : Optional[int]=0.0 , A_ : str=0.0 , A_ : Dict=0.02 , A_ : Union[str, Any]=1E-12 , A_ : Optional[int]=224 , A_ : int=16 , A_ : Optional[Any]=3 , A_ : Optional[int]=True , A_ : Dict=16 , **A_ : str , ) -> Tuple: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = initializer_range lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = qkv_bias lowerCamelCase_ = encoder_stride class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = version.parse('''1.11''' ) @property def a__ ( self : Optional[Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def a__ ( self : Optional[int] ) -> float: """simple docstring""" return 1E-4
208
def _SCREAMING_SNAKE_CASE ( lowercase : Tuple , lowercase : Dict , lowercase : List[str] , lowercase : Dict , lowercase : Dict , lowercase : List[str] ): '''simple docstring''' if index == r: for j in range(lowercase ): print(data[j] , end=' ' ) print(' ' ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location lowerCamelCase_ = arr[i] combination_util(lowercase , lowercase , lowercase , index + 1 , lowercase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(lowercase , lowercase , lowercase , lowercase , lowercase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def _SCREAMING_SNAKE_CASE ( lowercase : int , lowercase : Any , lowercase : Tuple ): '''simple docstring''' lowerCamelCase_ = [0] * r # Print all combination using temporary array 'data[]' combination_util(lowercase , lowercase , lowercase , 0 , lowercase , 0 ) if __name__ == "__main__": # Driver code to check the function above lowerCamelCase : int = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
208
1
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' @parameterized.expand([(None,), ("""foo.json""",)] ) def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Optional[Any] = GenerationConfig( do_sample=lowerCamelCase__, temperature=0.7, length_penalty=1.0, bad_words_ids=[[1, 2, 3], [4, 5]], ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__, config_name=lowerCamelCase__ ) A : Any = GenerationConfig.from_pretrained(lowerCamelCase__, config_name=lowerCamelCase__ ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample, lowerCamelCase__ ) self.assertEqual(loaded_config.temperature, 0.7 ) self.assertEqual(loaded_config.length_penalty, 1.0 ) self.assertEqual(loaded_config.bad_words_ids, [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k, 50 ) self.assertEqual(loaded_config.max_length, 20 ) self.assertEqual(loaded_config.max_time, lowerCamelCase__ ) def _lowerCAmelCase ( self ): A : Union[str, Any] = AutoConfig.from_pretrained("""gpt2""" ) A : str = GenerationConfig.from_model_config(lowerCamelCase__ ) A : Optional[Any] = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(lowerCamelCase__, lowerCamelCase__ ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id, default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id, model_config.eos_token_id ) def _lowerCAmelCase ( self ): A : List[Any] = GenerationConfig() A : List[Any] = { """max_new_tokens""": 1024, """foo""": """bar""", } A : int = copy.deepcopy(lowerCamelCase__ ) A : List[str] = generation_config.update(**lowerCamelCase__ ) # update_kwargs was not modified (no side effects) self.assertEqual(lowerCamelCase__, lowerCamelCase__ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens, 1024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(lowerCamelCase__, {"""foo""": """bar"""} ) def _lowerCAmelCase ( self ): A : Optional[Any] = GenerationConfig() A : Optional[int] = """bar""" with tempfile.TemporaryDirectory("""test-generation-config""" ) as tmp_dir: generation_config.save_pretrained(lowerCamelCase__ ) A : Any = GenerationConfig.from_pretrained(lowerCamelCase__ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo, """bar""" ) A : str = GenerationConfig.from_model_config(lowerCamelCase__ ) assert not hasattr(lowerCamelCase__, """foo""" ) # no new kwargs should be initialized if from config def _lowerCAmelCase ( self ): A : List[Any] = GenerationConfig() self.assertEqual(default_config.temperature, 1.0 ) self.assertEqual(default_config.do_sample, lowerCamelCase__ ) self.assertEqual(default_config.num_beams, 1 ) A : List[str] = GenerationConfig( do_sample=lowerCamelCase__, temperature=0.7, length_penalty=1.0, bad_words_ids=[[1, 2, 3], [4, 5]], ) self.assertEqual(config.temperature, 0.7 ) self.assertEqual(config.do_sample, lowerCamelCase__ ) self.assertEqual(config.num_beams, 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ ) A : Union[str, Any] = GenerationConfig.from_pretrained(lowerCamelCase__, temperature=1.0 ) self.assertEqual(loaded_config.temperature, 1.0 ) self.assertEqual(loaded_config.do_sample, lowerCamelCase__ ) self.assertEqual(loaded_config.num_beams, 1 ) # default value @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' @classmethod def _lowerCAmelCase ( cls ): A : Any = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def _lowerCAmelCase ( cls ): try: delete_repo(token=cls._token, repo_id="""test-generation-config""" ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id="""valid_org/test-generation-config-org""" ) except HTTPError: pass def _lowerCAmelCase ( self ): A : Any = GenerationConfig( do_sample=lowerCamelCase__, temperature=0.7, length_penalty=1.0, ) config.push_to_hub("""test-generation-config""", use_auth_token=self._token ) A : List[Any] = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__, getattr(lowerCamelCase__, lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token, repo_id="""test-generation-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__, repo_id="""test-generation-config""", push_to_hub=lowerCamelCase__, use_auth_token=self._token ) A : str = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__, getattr(lowerCamelCase__, lowerCamelCase__ ) ) def _lowerCAmelCase ( self ): A : List[Any] = GenerationConfig( do_sample=lowerCamelCase__, temperature=0.7, length_penalty=1.0, ) config.push_to_hub("""valid_org/test-generation-config-org""", use_auth_token=self._token ) A : Optional[Any] = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__, getattr(lowerCamelCase__, lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token, repo_id="""valid_org/test-generation-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__, repo_id="""valid_org/test-generation-config-org""", push_to_hub=lowerCamelCase__, use_auth_token=self._token ) A : Tuple = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__, getattr(lowerCamelCase__, lowerCamelCase__ ) )
116
'''simple docstring''' import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def lowerCamelCase ( __lowerCamelCase : Tuple ) ->Tuple: _SCREAMING_SNAKE_CASE = fname.split(os.path.sep )[-1] return re.search(R"""^(.*)_\d+\.jpg$""" , __lowerCamelCase ).groups()[0] class a_ ( snake_case_ ): '''simple docstring''' def __init__( self , A , A=None , A=None ) -> int: _SCREAMING_SNAKE_CASE = file_names _SCREAMING_SNAKE_CASE = image_transform _SCREAMING_SNAKE_CASE = label_to_id def __len__( self ) -> Optional[Any]: return len(self.file_names ) def __getitem__( self , A ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = self.file_names[idx] _SCREAMING_SNAKE_CASE = PIL.Image.open(A ) _SCREAMING_SNAKE_CASE = raw_image.convert("""RGB""" ) if self.image_transform is not None: _SCREAMING_SNAKE_CASE = self.image_transform(A ) _SCREAMING_SNAKE_CASE = extract_label(A ) if self.label_to_id is not None: _SCREAMING_SNAKE_CASE = self.label_to_id[label] return {"image": image, "label": label} def lowerCamelCase ( __lowerCamelCase : Any , __lowerCamelCase : Tuple ) ->str: # Initialize accelerator if args.with_tracking: _SCREAMING_SNAKE_CASE = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="""all""" , project_dir=args.project_dir ) else: _SCREAMING_SNAKE_CASE = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _SCREAMING_SNAKE_CASE = config["""lr"""] _SCREAMING_SNAKE_CASE = int(config["""num_epochs"""] ) _SCREAMING_SNAKE_CASE = int(config["""seed"""] ) _SCREAMING_SNAKE_CASE = int(config["""batch_size"""] ) _SCREAMING_SNAKE_CASE = config["""image_size"""] if not isinstance(__lowerCamelCase , (list, tuple) ): _SCREAMING_SNAKE_CASE = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , """isdigit""" ): if args.checkpointing_steps == "epoch": _SCREAMING_SNAKE_CASE = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): _SCREAMING_SNAKE_CASE = int(args.checkpointing_steps ) else: raise ValueError( F'Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.' ) else: _SCREAMING_SNAKE_CASE = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: _SCREAMING_SNAKE_CASE = os.path.split(__lowerCamelCase )[-1].split(""".""" )[0] accelerator.init_trackers(__lowerCamelCase , __lowerCamelCase ) # Grab all the image filenames _SCREAMING_SNAKE_CASE = [os.path.join(args.data_dir , __lowerCamelCase ) for fname in os.listdir(args.data_dir ) if fname.endswith(""".jpg""" )] # Build the label correspondences _SCREAMING_SNAKE_CASE = [extract_label(__lowerCamelCase ) for fname in file_names] _SCREAMING_SNAKE_CASE = list(set(__lowerCamelCase ) ) id_to_label.sort() _SCREAMING_SNAKE_CASE = {lbl: i for i, lbl in enumerate(__lowerCamelCase )} # Set the seed before splitting the data. np.random.seed(__lowerCamelCase ) torch.manual_seed(__lowerCamelCase ) torch.cuda.manual_seed_all(__lowerCamelCase ) # Split our filenames between train and validation _SCREAMING_SNAKE_CASE = np.random.permutation(len(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = int(0.8 * len(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = random_perm[:cut] _SCREAMING_SNAKE_CASE = random_perm[cut:] # For training we use a simple RandomResizedCrop _SCREAMING_SNAKE_CASE = Compose([RandomResizedCrop(__lowerCamelCase , scale=(0.5, 1.0) ), ToTensor()] ) _SCREAMING_SNAKE_CASE = PetsDataset( [file_names[i] for i in train_split] , image_transform=__lowerCamelCase , label_to_id=__lowerCamelCase ) # For evaluation, we use a deterministic Resize _SCREAMING_SNAKE_CASE = Compose([Resize(__lowerCamelCase ), ToTensor()] ) _SCREAMING_SNAKE_CASE = PetsDataset([file_names[i] for i in eval_split] , image_transform=__lowerCamelCase , label_to_id=__lowerCamelCase ) # Instantiate dataloaders. _SCREAMING_SNAKE_CASE = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) _SCREAMING_SNAKE_CASE = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _SCREAMING_SNAKE_CASE = create_model("""resnet50d""" , pretrained=__lowerCamelCase , num_classes=len(__lowerCamelCase ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _SCREAMING_SNAKE_CASE = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): _SCREAMING_SNAKE_CASE = False for param in model.get_classifier().parameters(): _SCREAMING_SNAKE_CASE = True # We normalize the batches of images to be a bit faster. _SCREAMING_SNAKE_CASE = torch.tensor(model.default_cfg["""mean"""] )[None, :, None, None].to(accelerator.device ) _SCREAMING_SNAKE_CASE = torch.tensor(model.default_cfg["""std"""] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer _SCREAMING_SNAKE_CASE = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler _SCREAMING_SNAKE_CASE = OneCycleLR(optimizer=__lowerCamelCase , max_lr=__lowerCamelCase , epochs=__lowerCamelCase , steps_per_epoch=len(__lowerCamelCase ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.prepare( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # We need to keep track of how many total steps we have iterated over _SCREAMING_SNAKE_CASE = 0 # We also need to keep track of the starting epoch so files are named properly _SCREAMING_SNAKE_CASE = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(F'Resumed from checkpoint: {args.resume_from_checkpoint}' ) accelerator.load_state(args.resume_from_checkpoint ) _SCREAMING_SNAKE_CASE = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint _SCREAMING_SNAKE_CASE = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) _SCREAMING_SNAKE_CASE = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` _SCREAMING_SNAKE_CASE = os.path.splitext(__lowerCamelCase )[0] if "epoch" in training_difference: _SCREAMING_SNAKE_CASE = int(training_difference.replace("""epoch_""" , """""" ) ) + 1 _SCREAMING_SNAKE_CASE = None else: _SCREAMING_SNAKE_CASE = int(training_difference.replace("""step_""" , """""" ) ) _SCREAMING_SNAKE_CASE = resume_step // len(__lowerCamelCase ) resume_step -= starting_epoch * len(__lowerCamelCase ) # Now we train the model for epoch in range(__lowerCamelCase , __lowerCamelCase ): model.train() if args.with_tracking: _SCREAMING_SNAKE_CASE = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step _SCREAMING_SNAKE_CASE = accelerator.skip_first_batches(__lowerCamelCase , __lowerCamelCase ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader _SCREAMING_SNAKE_CASE = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. _SCREAMING_SNAKE_CASE = {k: v.to(accelerator.device ) for k, v in batch.items()} _SCREAMING_SNAKE_CASE = (batch["""image"""] - mean) / std _SCREAMING_SNAKE_CASE = model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = torch.nn.functional.cross_entropy(__lowerCamelCase , batch["""label"""] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(__lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(__lowerCamelCase , __lowerCamelCase ): _SCREAMING_SNAKE_CASE = F'step_{overall_step}' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir , __lowerCamelCase ) accelerator.save_state(__lowerCamelCase ) model.eval() _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. _SCREAMING_SNAKE_CASE = {k: v.to(accelerator.device ) for k, v in batch.items()} _SCREAMING_SNAKE_CASE = (batch["""image"""] - mean) / std with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = outputs.argmax(dim=-1 ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.gather_for_metrics((predictions, batch["""label"""]) ) _SCREAMING_SNAKE_CASE = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() _SCREAMING_SNAKE_CASE = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}: {100 * eval_metric:.2f}' ) if args.with_tracking: accelerator.log( { """accuracy""": 100 * eval_metric, """train_loss""": total_loss.item() / len(__lowerCamelCase ), """epoch""": epoch, } , step=__lowerCamelCase , ) if checkpointing_steps == "epoch": _SCREAMING_SNAKE_CASE = F'epoch_{epoch}' if args.output_dir is not None: _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir , __lowerCamelCase ) accelerator.save_state(__lowerCamelCase ) if args.with_tracking: accelerator.end_training() def lowerCamelCase ( ) ->int: _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument("""--data_dir""" , required=__lowerCamelCase , help="""The data folder on disk.""" ) parser.add_argument("""--fp16""" , action="""store_true""" , help="""If passed, will use FP16 training.""" ) parser.add_argument( """--mixed_precision""" , type=__lowerCamelCase , default=__lowerCamelCase , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) parser.add_argument( """--checkpointing_steps""" , type=__lowerCamelCase , default=__lowerCamelCase , help="""Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch.""" , ) parser.add_argument( """--output_dir""" , type=__lowerCamelCase , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--resume_from_checkpoint""" , type=__lowerCamelCase , default=__lowerCamelCase , help="""If the training should continue from a checkpoint folder.""" , ) parser.add_argument( """--with_tracking""" , action="""store_true""" , help="""Whether to load in all available experiment trackers from the environment and use them for logging.""" , ) parser.add_argument( """--project_dir""" , type=__lowerCamelCase , default="""logs""" , help="""Location on where to store experiment tracking logs` and relevent project information""" , ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = {"""lr""": 3e-2, """num_epochs""": 3, """seed""": 42, """batch_size""": 64, """image_size""": 224} training_function(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
58
0
'''simple docstring''' import gc import threading import time import psutil import torch class __magic_name__ : def __init__( self : int ): _a : Optional[int] = psutil.Process() _a : str = False def __lowercase ( self : Optional[Any] ): _a : str = -1 while True: _a : Dict = max(self.process.memory_info().rss ,self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def __lowercase ( self : List[str] ): _a : List[str] = True _a : Tuple = threading.Thread(target=self.peak_monitor ) _a : Optional[Any] = True self.thread.start() def __lowercase ( self : Union[str, Any] ): _a : Any = False self.thread.join() return self.cpu_memory_peak __lowerCAmelCase = PeakCPUMemory() def __lowerCamelCase ( ) -> Tuple: # Time _a : Optional[int] = {'time': time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem _a : Optional[int] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): _a : Union[str, Any] = torch.cuda.memory_allocated(lowerCAmelCase_ ) torch.cuda.reset_peak_memory_stats() return measures def __lowerCamelCase ( lowerCAmelCase_ ) -> List[Any]: # Time _a : Any = {'time': time.time() - start_measures['time']} gc.collect() torch.cuda.empty_cache() # CPU mem _a : List[str] = (psutil.Process().memory_info().rss - start_measures['cpu']) / 2**20 _a : Union[str, Any] = (cpu_peak_tracker.stop() - start_measures['cpu']) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): _a : int = (torch.cuda.memory_allocated(lowerCAmelCase_ ) - start_measures[str(lowerCAmelCase_ )]) / 2**20 _a : List[str] = (torch.cuda.max_memory_allocated(lowerCAmelCase_ ) - start_measures[str(lowerCAmelCase_ )]) / 2**20 return measures def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: print(f"""{description}:""" ) print(f"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(f"""- GPU {i} allocated: {measures[str(lowerCAmelCase_ )]:.2f}MiB""" ) _a : Union[str, Any] = measures[f"""{i}-peak"""] print(f"""- GPU {i} peak: {peak:.2f}MiB""" ) print(f"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(f"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
350
'''simple docstring''' import heapq def __lowerCamelCase ( lowerCAmelCase_ ) -> set[int]: _a : list[list] = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(lowerCAmelCase_ , [-1 * len(lowerCAmelCase_ ), (key, value)] ) # chosen_vertices = set of chosen vertices _a : str = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _a : Optional[int] = heapq.heappop(lowerCAmelCase_ )[1][0] chosen_vertices.add(lowerCAmelCase_ ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _a : Dict = elem[1][1].index(lowerCAmelCase_ ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(lowerCAmelCase_ ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
107
0
'''simple docstring''' import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class a_ ( _lowerCAmelCase ): def __init__( self : Optional[int] , lowercase : List[str]=0.01 , lowercase : Dict=1_000 ): """simple docstring""" lowercase_ :int = p_stop lowercase_ :Dict = max_length def __iter__( self : Dict ): """simple docstring""" lowercase_ :Union[str, Any] = 0 lowercase_ :Optional[Any] = False while not stop and count < self.max_length: yield count count += 1 lowercase_ :Union[str, Any] = random.random() < self.p_stop class a_ ( unittest.TestCase ): def lowercase__ ( self : Tuple , lowercase : Union[str, Any] , lowercase : int , lowercase : Union[str, Any]=False , lowercase : Union[str, Any]=True ): """simple docstring""" lowercase_ :Optional[int] = [ BatchSamplerShard(lowercase , 2 , lowercase , split_batches=lowercase , even_batches=lowercase ) for i in range(2 ) ] lowercase_ :Optional[Any] = [list(lowercase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowercase ) for shard in batch_sampler_shards] , [len(lowercase ) for e in expected] ) self.assertListEqual(lowercase , lowercase ) def lowercase__ ( self : Dict ): """simple docstring""" lowercase_ :Dict = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :List[str] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowercase , lowercase ) lowercase_ :str = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowercase_ :int = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :List[str] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(lowercase , lowercase ) lowercase_ :Optional[int] = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowercase_ :Tuple = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(lowercase , lowercase ) lowercase_ :Any = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowercase_ :List[str] = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Tuple = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(lowercase , lowercase ) lowercase_ :List[str] = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase ) # Check the shards when the dataset is very small. lowercase_ :Optional[int] = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Optional[Any] = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowercase , lowercase ) lowercase_ :Any = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :int = [[], []] self.check_batch_sampler_shards(lowercase , lowercase ) def lowercase__ ( self : Dict ): """simple docstring""" lowercase_ :Union[str, Any] = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) lowercase_ :Optional[int] = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size. lowercase_ :Union[str, Any] = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :List[str] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) lowercase_ :Tuple = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowercase_ :Tuple = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :int = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) lowercase_ :List[Any] = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Optional[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) # Check the shards when the dataset is very small. lowercase_ :List[Any] = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) lowercase_ :List[str] = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Dict = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase ) def lowercase__ ( self : Dict ): """simple docstring""" lowercase_ :Tuple = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :List[str] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) lowercase_ :Optional[Any] = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowercase_ :Dict = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Dict = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) lowercase_ :List[str] = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowercase_ :Union[str, Any] = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :str = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) lowercase_ :List[Any] = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :int = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowercase_ :int = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) lowercase_ :List[Any] = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :str = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) # Check the shards when the dataset is very small. lowercase_ :str = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :Optional[int] = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) lowercase_ :Optional[Any] = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase ) lowercase_ :List[Any] = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , even_batches=lowercase ) def lowercase__ ( self : int ): """simple docstring""" lowercase_ :Union[str, Any] = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Tuple = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) lowercase_ :Dict = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowercase ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size. lowercase_ :Tuple = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Any = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) lowercase_ :str = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowercase_ :List[str] = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) lowercase_ :Dict = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) # Check the shards when the dataset is very small. lowercase_ :Optional[int] = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Optional[Any] = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) lowercase_ :Tuple = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :Dict = [[], []] self.check_batch_sampler_shards(lowercase , lowercase , split_batches=lowercase , even_batches=lowercase ) def lowercase__ ( self : Optional[Any] ): """simple docstring""" lowercase_ :str = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] lowercase_ :Tuple = [BatchSamplerShard(lowercase , 2 , lowercase , even_batches=lowercase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def lowercase__ ( self : Union[str, Any] , lowercase : Union[str, Any] , lowercase : Dict , lowercase : Tuple , lowercase : List[Any]=False , lowercase : int=2 , lowercase : List[str]=False ): """simple docstring""" random.seed(lowercase ) lowercase_ :Optional[int] = list(lowercase ) lowercase_ :Optional[Any] = [ IterableDatasetShard( lowercase , batch_size=lowercase , drop_last=lowercase , num_processes=lowercase , process_index=lowercase , split_batches=lowercase , ) for i in range(lowercase ) ] lowercase_ :str = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowercase ) iterable_dataset_lists.append(list(lowercase ) ) lowercase_ :Dict = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size lowercase_ :Dict = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowercase ) , len(lowercase ) ) self.assertTrue(len(lowercase ) % shard_batch_size == 0 ) lowercase_ :int = [] for idx in range(0 , len(lowercase ) , lowercase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowercase ) < len(lowercase ): reference += reference self.assertListEqual(lowercase , reference[: len(lowercase )] ) def lowercase__ ( self : int ): """simple docstring""" lowercase_ :str = 42 lowercase_ :Any = RandomIterableDataset() self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) # Edge case with a very small dataset lowercase_ :List[Any] = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) self.check_iterable_dataset_shards(lowercase , lowercase , batch_size=4 , drop_last=lowercase , split_batches=lowercase ) def lowercase__ ( self : Optional[Any] ): """simple docstring""" lowercase_ :Any = BatchSampler(range(16 ) , batch_size=4 , drop_last=lowercase ) lowercase_ :List[Any] = SkipBatchSampler(lowercase , 2 ) self.assertListEqual(list(lowercase ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase__ ( self : Tuple ): """simple docstring""" lowercase_ :Any = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase__ ( self : Union[str, Any] ): """simple docstring""" lowercase_ :Tuple = DataLoader(list(range(16 ) ) , batch_size=4 ) lowercase_ :Optional[int] = skip_first_batches(lowercase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase__ ( self : Optional[Any] ): """simple docstring""" lowercase_ :Optional[Any] = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def lowercase__ ( self : Union[str, Any] ): """simple docstring""" Accelerator() lowercase_ :Union[str, Any] = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
223
'''simple docstring''' lowerCAmelCase : str =''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' lowerCAmelCase : int =[{'''type''': '''code''', '''content''': INSTALL_CONTENT}] lowerCAmelCase : List[str] ={ '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
223
1
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class __magic_name__ ( unittest.TestCase ): def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = { "task_specific_params": { "summarization": {"length_penalty": 1.0, "max_length": 128, "min_length": 12, "num_beams": 4}, "summarization_cnn": {"length_penalty": 2.0, "max_length": 142, "min_length": 56, "num_beams": 4}, "summarization_xsum": {"length_penalty": 1.0, "max_length": 62, "min_length": 11, "num_beams": 6}, } } UpperCamelCase_ = { "task_specific_params.summarization.length_penalty": 1.0, "task_specific_params.summarization.max_length": 128, "task_specific_params.summarization.min_length": 12, "task_specific_params.summarization.num_beams": 4, "task_specific_params.summarization_cnn.length_penalty": 2.0, "task_specific_params.summarization_cnn.max_length": 142, "task_specific_params.summarization_cnn.min_length": 56, "task_specific_params.summarization_cnn.num_beams": 4, "task_specific_params.summarization_xsum.length_penalty": 1.0, "task_specific_params.summarization_xsum.max_length": 62, "task_specific_params.summarization_xsum.min_length": 11, "task_specific_params.summarization_xsum.num_beams": 6, } self.assertEqual(flatten_dict(_lowercase ) , _lowercase ) def UpperCAmelCase_ ( self )-> Tuple: UpperCamelCase_ = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(_lowercase ) , x.transpose() ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(_lowercase , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase ) , transpose(_lowercase ).numpy() ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase , axes=(1, 2, 0) ) , transpose(_lowercase , axes=(1, 2, 0) ).numpy() ) ) @require_tf def UpperCAmelCase_ ( self )-> Tuple: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase ) , transpose(_lowercase ).numpy() ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase , axes=(1, 2, 0) ) , transpose(_lowercase , axes=(1, 2, 0) ).numpy() ) ) @require_flax def UpperCAmelCase_ ( self )-> str: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase ) , np.asarray(transpose(_lowercase ) ) ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(transpose(_lowercase , axes=(1, 2, 0) ) , np.asarray(transpose(_lowercase , axes=(1, 2, 0) ) ) ) ) def UpperCAmelCase_ ( self )-> Any: UpperCamelCase_ = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(_lowercase , (4, 3) ) , np.reshape(_lowercase , (4, 3) ) ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(_lowercase , (12, 5) ) , np.reshape(_lowercase , (12, 5) ) ) ) @require_torch def UpperCAmelCase_ ( self )-> Any: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (4, 3) ) , reshape(_lowercase , (4, 3) ).numpy() ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (12, 5) ) , reshape(_lowercase , (12, 5) ).numpy() ) ) @require_tf def UpperCAmelCase_ ( self )-> Dict: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (4, 3) ) , reshape(_lowercase , (4, 3) ).numpy() ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (12, 5) ) , reshape(_lowercase , (12, 5) ).numpy() ) ) @require_flax def UpperCAmelCase_ ( self )-> List[str]: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (4, 3) ) , np.asarray(reshape(_lowercase , (4, 3) ) ) ) ) UpperCamelCase_ = np.random.randn(3 , 4 , 5 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(reshape(_lowercase , (12, 5) ) , np.asarray(reshape(_lowercase , (12, 5) ) ) ) ) def UpperCAmelCase_ ( self )-> Tuple: UpperCamelCase_ = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(_lowercase ) , np.squeeze(_lowercase ) ) ) UpperCamelCase_ = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(_lowercase , axis=2 ) , np.squeeze(_lowercase , axis=2 ) ) ) @require_torch def UpperCAmelCase_ ( self )-> int: UpperCamelCase_ = np.random.randn(1 , 3 , 4 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase ) , squeeze(_lowercase ).numpy() ) ) UpperCamelCase_ = np.random.randn(1 , 4 , 1 , 5 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase , axis=2 ) , squeeze(_lowercase , axis=2 ).numpy() ) ) @require_tf def UpperCAmelCase_ ( self )-> str: UpperCamelCase_ = np.random.randn(1 , 3 , 4 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase ) , squeeze(_lowercase ).numpy() ) ) UpperCamelCase_ = np.random.randn(1 , 4 , 1 , 5 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase , axis=2 ) , squeeze(_lowercase , axis=2 ).numpy() ) ) @require_flax def UpperCAmelCase_ ( self )-> Any: UpperCamelCase_ = np.random.randn(1 , 3 , 4 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase ) , np.asarray(squeeze(_lowercase ) ) ) ) UpperCamelCase_ = np.random.randn(1 , 4 , 1 , 5 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(squeeze(_lowercase , axis=2 ) , np.asarray(squeeze(_lowercase , axis=2 ) ) ) ) def UpperCAmelCase_ ( self )-> List[str]: UpperCamelCase_ = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(_lowercase , axis=1 ) , np.expand_dims(_lowercase , axis=1 ) ) ) @require_torch def UpperCAmelCase_ ( self )-> List[Any]: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = torch.tensor(_lowercase ) self.assertTrue(np.allclose(expand_dims(_lowercase , axis=1 ) , expand_dims(_lowercase , axis=1 ).numpy() ) ) @require_tf def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = tf.constant(_lowercase ) self.assertTrue(np.allclose(expand_dims(_lowercase , axis=1 ) , expand_dims(_lowercase , axis=1 ).numpy() ) ) @require_flax def UpperCAmelCase_ ( self )-> Union[str, Any]: UpperCamelCase_ = np.random.randn(3 , 4 ) UpperCamelCase_ = jnp.array(_lowercase ) self.assertTrue(np.allclose(expand_dims(_lowercase , axis=1 ) , np.asarray(expand_dims(_lowercase , axis=1 ) ) ) )
60
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Union[str, Any] = { """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class __magic_name__ ( snake_case ): UpperCamelCase_ :List[Any] = """bridgetower_vision_model""" def __init__( self , _lowercase=768 , _lowercase=12 , _lowercase=3 , _lowercase=16 , _lowercase=288 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=False , _lowercase=True , _lowercase=False , **_lowercase , )-> Optional[Any]: super().__init__(**_lowercase ) UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_channels UpperCamelCase_ = patch_size UpperCamelCase_ = image_size UpperCamelCase_ = initializer_factor UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = stop_gradient UpperCamelCase_ = share_layernorm UpperCamelCase_ = remove_last_layer @classmethod def UpperCAmelCase_ ( cls , _lowercase , **_lowercase )-> "PretrainedConfig": UpperCamelCase_ , UpperCamelCase_ = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get("model_type" ) == "bridgetower": UpperCamelCase_ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_lowercase , **_lowercase ) class __magic_name__ ( snake_case ): UpperCamelCase_ :Optional[int] = """bridgetower_text_model""" def __init__( self , _lowercase=50_265 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=1 , _lowercase=3_072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=514 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase="absolute" , _lowercase=True , **_lowercase , )-> Optional[int]: super().__init__(**_lowercase ) UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = hidden_act UpperCamelCase_ = initializer_factor UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = type_vocab_size UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = position_embedding_type UpperCamelCase_ = use_cache UpperCamelCase_ = pad_token_id UpperCamelCase_ = bos_token_id UpperCamelCase_ = eos_token_id @classmethod def UpperCAmelCase_ ( cls , _lowercase , **_lowercase )-> "PretrainedConfig": UpperCamelCase_ , UpperCamelCase_ = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get("model_type" ) == "bridgetower": UpperCamelCase_ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_lowercase , **_lowercase ) class __magic_name__ ( snake_case ): UpperCamelCase_ :List[Any] = """bridgetower""" def __init__( self , _lowercase=True , _lowercase="gelu" , _lowercase=768 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=False , _lowercase="add" , _lowercase=12 , _lowercase=6 , _lowercase=False , _lowercase=False , _lowercase=None , _lowercase=None , **_lowercase , )-> List[Any]: # TODO: remove this once the Hub files are updated. UpperCamelCase_ = kwargs.pop("text_config_dict" , _lowercase ) UpperCamelCase_ = kwargs.pop("vision_config_dict" , _lowercase ) super().__init__(**_lowercase ) UpperCamelCase_ = share_cross_modal_transformer_layers UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_size UpperCamelCase_ = initializer_factor UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = share_link_tower_layers UpperCamelCase_ = link_tower_type UpperCamelCase_ = num_attention_heads UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = tie_word_embeddings UpperCamelCase_ = init_layernorm_from_vision_encoder if text_config is None: UpperCamelCase_ = {} logger.info("`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values." ) if vision_config is None: UpperCamelCase_ = {} logger.info("`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values." ) UpperCamelCase_ = BridgeTowerTextConfig(**_lowercase ) UpperCamelCase_ = BridgeTowerVisionConfig(**_lowercase ) @classmethod def UpperCAmelCase_ ( cls , _lowercase , _lowercase , **_lowercase )-> List[str]: return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_lowercase ) def UpperCAmelCase_ ( self )-> Union[str, Any]: UpperCamelCase_ = copy.deepcopy(self.__dict__ ) UpperCamelCase_ = self.text_config.to_dict() UpperCamelCase_ = self.vision_config.to_dict() UpperCamelCase_ = self.__class__.model_type return output
60
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL lowerCAmelCase__ : Optional[Any] = logging.get_logger(__name__) def a_ ( lowerCamelCase ): if isinstance(lowerCamelCase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowerCamelCase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowerCamelCase ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = ["pixel_values"] def __init__( self : Dict ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = PILImageResampling.BILINEAR ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Union[int, float] = 1 / 255 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,**lowerCamelCase__ : List[Any] ,): super().__init__(**lowerCamelCase__ ) UpperCAmelCase__ = size if size is not None else {'shortest_edge': 256} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ) UpperCAmelCase__ = crop_size if crop_size is not None else {'height': 224, 'width': 224} UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='crop_size' ) UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = do_center_crop UpperCAmelCase__ = crop_size UpperCAmelCase__ = resample UpperCAmelCase__ = do_rescale UpperCAmelCase__ = rescale_factor UpperCAmelCase__ = offset UpperCAmelCase__ = do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowerCAmelCase ( self : int ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : PILImageResampling = PILImageResampling.BILINEAR ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : List[str] ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ) if "shortest_edge" in size: UpperCAmelCase__ = get_resize_output_image_size(lowerCamelCase__ ,size['shortest_edge'] ,default_to_square=lowerCamelCase__ ) elif "height" in size and "width" in size: UpperCAmelCase__ = (size['height'], size['width']) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Dict[str, int] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : List[Any] ,): UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCamelCase__ ,size=(size['height'], size['width']) ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[int, float] ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Union[str, Any] ,): UpperCAmelCase__ = image.astype(np.floataa ) if offset: UpperCAmelCase__ = image - (scale / 2) return rescale(lowerCamelCase__ ,scale=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : np.ndarray ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Union[float, List[float]] ,lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase__ : Dict ,): return normalize(lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ,data_format=lowerCamelCase__ ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : ImageInput ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : float = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[ChannelDimension] = ChannelDimension.FIRST ,): if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) if offset and not do_rescale: raise ValueError('For offset, do_rescale must also be set to True.' ) # All transformations expect numpy arrays. UpperCAmelCase__ = to_numpy_array(lowerCamelCase__ ) if do_resize: UpperCAmelCase__ = self.resize(image=lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ) if do_center_crop: UpperCAmelCase__ = self.center_crop(lowerCamelCase__ ,size=lowerCamelCase__ ) if do_rescale: UpperCAmelCase__ = self.rescale(image=lowerCamelCase__ ,scale=lowerCamelCase__ ,offset=lowerCamelCase__ ) if do_normalize: UpperCAmelCase__ = self.normalize(image=lowerCamelCase__ ,mean=lowerCamelCase__ ,std=lowerCamelCase__ ) UpperCAmelCase__ = to_channel_dimension_format(lowerCamelCase__ ,lowerCamelCase__ ) return image def __lowerCAmelCase ( self : Tuple ,lowerCamelCase__ : ImageInput ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : PILImageResampling = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Dict[str, int] = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : float = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : bool = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[float, List[float]]] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,lowerCamelCase__ : ChannelDimension = ChannelDimension.FIRST ,**lowerCamelCase__ : Dict ,): UpperCAmelCase__ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ = resample if resample is not None else self.resample UpperCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ = offset if offset is not None else self.offset UpperCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase__ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase__ = image_std if image_std is not None else self.image_std UpperCAmelCase__ = size if size is not None else self.size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,default_to_square=lowerCamelCase__ ) UpperCAmelCase__ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ = get_size_dict(lowerCamelCase__ ,param_name='crop_size' ) if not valid_images(lowerCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) UpperCAmelCase__ = make_batched(lowerCamelCase__ ) UpperCAmelCase__ = [ [ self._preprocess_image( image=lowerCamelCase__ ,do_resize=lowerCamelCase__ ,size=lowerCamelCase__ ,resample=lowerCamelCase__ ,do_center_crop=lowerCamelCase__ ,crop_size=lowerCamelCase__ ,do_rescale=lowerCamelCase__ ,rescale_factor=lowerCamelCase__ ,offset=lowerCamelCase__ ,do_normalize=lowerCamelCase__ ,image_mean=lowerCamelCase__ ,image_std=lowerCamelCase__ ,data_format=lowerCamelCase__ ,) for img in video ] for video in videos ] UpperCAmelCase__ = {'pixel_values': videos} return BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ )
98
'''simple docstring''' from datetime import datetime import requests def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase ='https://downloadgram.net/wp-json/wppress/video-downloader/video?url=' __lowercase =requests.get(base_url + url ).json()[0]['urls'][0]['src'] return requests.get(_lowerCAmelCase ).content if __name__ == "__main__": lowerCamelCase = input("""Enter Video/IGTV url: """).strip() lowerCamelCase = f"{datetime.now():%Y-%m-%d_%H:%M:%S}.mp4" with open(file_name, """wb""") as fp: fp.write(download_video(url)) print(f"Done. Video saved to disk as {file_name}.")
166
0
from __future__ import annotations class lowerCamelCase__ : '''simple docstring''' def __init__(self ,__lowerCamelCase ) -> None: """simple docstring""" lowerCAmelCase__ : str = order # a_{0} ... a_{k} lowerCAmelCase__ : int = [1.0] + [0.0] * order # b_{0} ... b_{k} lowerCAmelCase__ : str = [1.0] + [0.0] * order # x[n-1] ... x[n-k] lowerCAmelCase__ : Any = [0.0] * self.order # y[n-1] ... y[n-k] lowerCAmelCase__ : Any = [0.0] * self.order def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase ) -> None: """simple docstring""" if len(lowerCAmelCase__ ) < self.order: lowerCAmelCase__ : Optional[int] = [1.0, *a_coeffs] if len(lowerCAmelCase__ ) != self.order + 1: lowerCAmelCase__ : Union[str, Any] = ( f"""Expected a_coeffs to have {self.order + 1} elements """ f"""for {self.order}-order filter, got {len(lowerCAmelCase__ )}""" ) raise ValueError(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) != self.order + 1: lowerCAmelCase__ : Any = ( f"""Expected b_coeffs to have {self.order + 1} elements """ f"""for {self.order}-order filter, got {len(lowerCAmelCase__ )}""" ) raise ValueError(lowerCAmelCase__ ) lowerCAmelCase__ : List[Any] = a_coeffs lowerCAmelCase__ : str = b_coeffs def lowerCAmelCase__ (self ,__lowerCamelCase ) -> float: """simple docstring""" lowerCAmelCase__ : Optional[Any] = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 ,self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) lowerCAmelCase__ : str = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] lowerCAmelCase__ : Any = self.input_history[:-1] lowerCAmelCase__ : Optional[Any] = self.output_history[:-1] lowerCAmelCase__ : List[str] = sample lowerCAmelCase__ : Tuple = result return result
366
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCamelCase__ ( lowerCamelCase__ , unittest.TestCase): '''simple docstring''' snake_case_ =KandinskyVaaPriorPipeline snake_case_ =["""prompt"""] snake_case_ =["""prompt""", """negative_prompt"""] snake_case_ =[ """num_images_per_prompt""", """generator""", """num_inference_steps""", """latents""", """negative_prompt""", """guidance_scale""", """output_type""", """return_dict""", ] snake_case_ =False @property def lowerCAmelCase__ (self ) -> Any: """simple docstring""" return 32 @property def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" return 32 @property def lowerCAmelCase__ (self ) -> Tuple: """simple docstring""" return self.time_input_dim @property def lowerCAmelCase__ (self ) -> Optional[int]: """simple docstring""" return self.time_input_dim * 4 @property def lowerCAmelCase__ (self ) -> Dict: """simple docstring""" return 1_00 @property def lowerCAmelCase__ (self ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Any = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def lowerCAmelCase__ (self ) -> str: """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ : str = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=self.text_embedder_hidden_size ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) return CLIPTextModelWithProjection(__lowerCamelCase ) @property def lowerCAmelCase__ (self ) -> int: """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = { '''num_attention_heads''': 2, '''attention_head_dim''': 12, '''embedding_dim''': self.text_embedder_hidden_size, '''num_layers''': 1, } lowerCAmelCase__ : List[str] = PriorTransformer(**__lowerCamelCase ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 lowerCAmelCase__ : int = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ : Any = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size ,image_size=2_24 ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=37 ,num_attention_heads=4 ,num_channels=3 ,num_hidden_layers=5 ,patch_size=14 ,) lowerCAmelCase__ : int = CLIPVisionModelWithProjection(__lowerCamelCase ) return model @property def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ : Optional[Any] = CLIPImageProcessor( crop_size=2_24 ,do_center_crop=__lowerCamelCase ,do_normalize=__lowerCamelCase ,do_resize=__lowerCamelCase ,image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] ,image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] ,resample=3 ,size=2_24 ,) return image_processor def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ : List[Any] = self.dummy_prior lowerCAmelCase__ : List[Any] = self.dummy_image_encoder lowerCAmelCase__ : Optional[Any] = self.dummy_text_encoder lowerCAmelCase__ : Optional[int] = self.dummy_tokenizer lowerCAmelCase__ : str = self.dummy_image_processor lowerCAmelCase__ : Union[str, Any] = UnCLIPScheduler( variance_type='''fixed_small_log''' ,prediction_type='''sample''' ,num_train_timesteps=10_00 ,clip_sample=__lowerCamelCase ,clip_sample_range=10.0 ,) lowerCAmelCase__ : int = { '''prior''': prior, '''image_encoder''': image_encoder, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''scheduler''': scheduler, '''image_processor''': image_processor, } return components def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase=0 ) -> int: """simple docstring""" if str(__lowerCamelCase ).startswith('''mps''' ): lowerCAmelCase__ : Dict = torch.manual_seed(__lowerCamelCase ) else: lowerCAmelCase__ : Any = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) lowerCAmelCase__ : Optional[Any] = { '''prompt''': '''horse''', '''generator''': generator, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" lowerCAmelCase__ : str = '''cpu''' lowerCAmelCase__ : Optional[int] = self.get_dummy_components() lowerCAmelCase__ : int = self.pipeline_class(**__lowerCamelCase ) lowerCAmelCase__ : int = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) lowerCAmelCase__ : List[Any] = pipe(**self.get_dummy_inputs(__lowerCamelCase ) ) lowerCAmelCase__ : Union[str, Any] = output.image_embeds lowerCAmelCase__ : Tuple = pipe( **self.get_dummy_inputs(__lowerCamelCase ) ,return_dict=__lowerCamelCase ,)[0] lowerCAmelCase__ : Union[str, Any] = image[0, -10:] lowerCAmelCase__ : str = image_from_tuple[0, -10:] assert image.shape == (1, 32) lowerCAmelCase__ : int = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase__ (self ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : List[Any] = torch_device == '''cpu''' lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Optional[int] = False self._test_inference_batch_single_identical( test_max_difference=__lowerCamelCase ,relax_max_difference=__lowerCamelCase ,test_mean_pixel_difference=__lowerCamelCase ,) @skip_mps def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ : int = torch_device == '''cpu''' lowerCAmelCase__ : int = False self._test_attention_slicing_forward_pass( test_max_difference=__lowerCamelCase ,test_mean_pixel_difference=__lowerCamelCase ,)
94
0
import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor a__: int = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowerCamelCase_ ): def __init__( self,*__lowerCamelCase,**__lowerCamelCase ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''',__SCREAMING_SNAKE_CASE,) super().__init__(*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE )
193
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase__ ( unittest.TestCase ): @slow def _snake_case ( self ): """simple docstring""" lowercase_ : Dict = AutoModelForSeqaSeqLM.from_pretrained('''google/mt5-small''' , return_dict=__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = AutoTokenizer.from_pretrained('''google/mt5-small''' ) lowercase_ : int = tokenizer('''Hello there''' , return_tensors='''pt''' ).input_ids lowercase_ : Union[str, Any] = tokenizer('''Hi I am''' , return_tensors='''pt''' ).input_ids lowercase_ : Union[str, Any] = model(input_ids.to(__SCREAMING_SNAKE_CASE ) , labels=labels.to(__SCREAMING_SNAKE_CASE ) ).loss lowercase_ : int = -(labels.shape[-1] * loss.item()) lowercase_ : Any = -84.9_127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
93
0
import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase : def __init__( self : List[Any] , _UpperCamelCase : int ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = data SCREAMING_SNAKE_CASE = [0x6_7_4_5_2_3_0_1, 0xE_F_C_D_A_B_8_9, 0x9_8_B_A_D_C_F_E, 0x1_0_3_2_5_4_7_6, 0xC_3_D_2_E_1_F_0] @staticmethod def __snake_case( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[Any] ) -> Tuple: '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0xF_F_F_F_F_F_F_F def __snake_case( self : List[Any] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = b"\x80" + b"\x00" * (63 - (len(self.data ) + 8) % 64) SCREAMING_SNAKE_CASE = self.data + padding + struct.pack(">Q" , 8 * len(self.data ) ) return padded_data def __snake_case( self : Optional[int] ) -> List[Any]: '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __snake_case( self : Optional[int] , _UpperCamelCase : Any ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = list(struct.unpack(">16L" , _UpperCamelCase ) ) + [0] * 64 for i in range(16 , 80 ): SCREAMING_SNAKE_CASE = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __snake_case( self : Tuple ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.padding() SCREAMING_SNAKE_CASE = self.split_blocks() for block in self.blocks: SCREAMING_SNAKE_CASE = self.expand_block(_UpperCamelCase ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.h for i in range(0 , 80 ): if 0 <= i < 20: SCREAMING_SNAKE_CASE = (b & c) | ((~b) & d) SCREAMING_SNAKE_CASE = 0x5_A_8_2_7_9_9_9 elif 20 <= i < 40: SCREAMING_SNAKE_CASE = b ^ c ^ d SCREAMING_SNAKE_CASE = 0x6_E_D_9_E_B_A_1 elif 40 <= i < 60: SCREAMING_SNAKE_CASE = (b & c) | (b & d) | (c & d) SCREAMING_SNAKE_CASE = 0x8_F_1_B_B_C_D_C elif 60 <= i < 80: SCREAMING_SNAKE_CASE = b ^ c ^ d SCREAMING_SNAKE_CASE = 0xC_A_6_2_C_1_D_6 SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = ( self.rotate(_UpperCamelCase , 5 ) + f + e + k + expanded_block[i] & 0xF_F_F_F_F_F_F_F, a, self.rotate(_UpperCamelCase , 30 ), c, d, ) SCREAMING_SNAKE_CASE = ( self.h[0] + a & 0xF_F_F_F_F_F_F_F, self.h[1] + b & 0xF_F_F_F_F_F_F_F, self.h[2] + c & 0xF_F_F_F_F_F_F_F, self.h[3] + d & 0xF_F_F_F_F_F_F_F, self.h[4] + e & 0xF_F_F_F_F_F_F_F, ) return ("{:08x}" * 5).format(*self.h ) def __lowerCamelCase (): SCREAMING_SNAKE_CASE = B"Test String" assert SHAaHash(UpperCAmelCase__ ).final_hash() == hashlib.shaa(UpperCAmelCase__ ).hexdigest() # noqa: S324 def __lowerCamelCase (): SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="Process some strings or files" ) parser.add_argument( "--string" , dest="input_string" , default="Hello World!! Welcome to Cryptography" , help="Hash the string" , ) parser.add_argument("--file" , dest="input_file" , help="Hash contents of a file" ) SCREAMING_SNAKE_CASE = parser.parse_args() SCREAMING_SNAKE_CASE = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , "rb" ) as f: SCREAMING_SNAKE_CASE = f.read() else: SCREAMING_SNAKE_CASE = bytes(UpperCAmelCase__ , "utf-8" ) print(SHAaHash(UpperCAmelCase__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
206
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( a ): lowercase__ : Dict = (UniPCMultistepScheduler,) lowercase__ : Optional[int] = (("""num_inference_steps""", 25),) def __snake_case( self : List[str] , **_UpperCamelCase : Tuple ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = { "num_train_timesteps": 1_000, "beta_start": 0.0_0_0_1, "beta_end": 0.0_2, "beta_schedule": "linear", "solver_order": 2, "solver_type": "bh2", } config.update(**_UpperCamelCase ) return config def __snake_case( self : List[str] , _UpperCamelCase : Dict=0 , **_UpperCamelCase : List[str] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class.from_pretrained(_UpperCamelCase ) new_scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals SCREAMING_SNAKE_CASE = dummy_past_residuals[: new_scheduler.config.solver_order] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = sample, sample for t in range(_UpperCamelCase , time_step + scheduler.config.solver_order + 1 ): SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = new_scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __snake_case( self : Any , _UpperCamelCase : Union[str, Any]=0 , **_UpperCamelCase : Dict ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residuals (must be after setting timesteps) SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class.from_pretrained(_UpperCamelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(_UpperCamelCase ) # copy over dummy past residual (must be after setting timesteps) SCREAMING_SNAKE_CASE = dummy_past_residuals[: new_scheduler.config.solver_order] SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = new_scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __snake_case( self : List[str] , _UpperCamelCase : Tuple=None , **_UpperCamelCase : List[Any] ) -> str: '''simple docstring''' if scheduler is None: SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = 10 SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter scheduler.set_timesteps(_UpperCamelCase ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE = model(_UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ).prev_sample return sample def __snake_case( self : int ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = dict(self.forward_default_kwargs ) SCREAMING_SNAKE_CASE = kwargs.pop("num_inference_steps" , _UpperCamelCase ) for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = self.dummy_sample SCREAMING_SNAKE_CASE = 0.1 * sample if num_inference_steps is not None and hasattr(_UpperCamelCase , "set_timesteps" ): scheduler.set_timesteps(_UpperCamelCase ) elif num_inference_steps is not None and not hasattr(_UpperCamelCase , "set_timesteps" ): SCREAMING_SNAKE_CASE = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) SCREAMING_SNAKE_CASE = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] SCREAMING_SNAKE_CASE = dummy_past_residuals[: scheduler.config.solver_order] SCREAMING_SNAKE_CASE = scheduler.timesteps[5] SCREAMING_SNAKE_CASE = scheduler.timesteps[6] SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __snake_case( self : str ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = UniPCMultistepScheduler(**self.get_scheduler_config() ) SCREAMING_SNAKE_CASE = self.full_loop(scheduler=_UpperCamelCase ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 SCREAMING_SNAKE_CASE = DPMSolverSinglestepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = DEISMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = UniPCMultistepScheduler.from_config(scheduler.config ) SCREAMING_SNAKE_CASE = self.full_loop(scheduler=_UpperCamelCase ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 def __snake_case( self : Optional[int] ) -> List[Any]: '''simple docstring''' for timesteps in [25, 50, 100, 999, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCamelCase ) def __snake_case( self : Tuple ) -> Union[str, Any]: '''simple docstring''' self.check_over_configs(thresholding=_UpperCamelCase ) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=_UpperCamelCase , prediction_type=_UpperCamelCase , sample_max_value=_UpperCamelCase , solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , ) def __snake_case( self : Tuple ) -> int: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCamelCase ) def __snake_case( self : Dict ) -> int: '''simple docstring''' for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , prediction_type=_UpperCamelCase , ) SCREAMING_SNAKE_CASE = self.full_loop( solver_order=_UpperCamelCase , solver_type=_UpperCamelCase , prediction_type=_UpperCamelCase , ) assert not torch.isnan(_UpperCamelCase ).any(), "Samples have nan numbers" def __snake_case( self : List[str] ) -> Tuple: '''simple docstring''' self.check_over_configs(lower_order_final=_UpperCamelCase ) self.check_over_configs(lower_order_final=_UpperCamelCase ) def __snake_case( self : Optional[int] ) -> List[Any]: '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1_000]: self.check_over_forward(num_inference_steps=_UpperCamelCase , time_step=0 ) def __snake_case( self : Tuple ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = self.full_loop() SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.2_4_6_4 ) < 1e-3 def __snake_case( self : List[str] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.full_loop(prediction_type="v_prediction" ) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_mean.item() - 0.1_0_1_4 ) < 1e-3 def __snake_case( self : int ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(thresholding=_UpperCamelCase , dynamic_thresholding_ratio=0 ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = 10 SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter.half() scheduler.set_timesteps(_UpperCamelCase ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE = model(_UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ).prev_sample assert sample.dtype == torch.floataa def __snake_case( self : List[str] , **_UpperCamelCase : Dict ) -> Union[str, Any]: '''simple docstring''' for scheduler_class in self.scheduler_classes: SCREAMING_SNAKE_CASE = self.get_scheduler_config(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = scheduler_class(**_UpperCamelCase ) scheduler.set_timesteps(scheduler.config.num_train_timesteps ) assert len(scheduler.timesteps.unique() ) == scheduler.num_inference_steps
206
1
"""simple docstring""" class _UpperCAmelCase : def __init__( self :Union[str, Any] , __UpperCamelCase :Dict ): A = len(__UpperCAmelCase ) A = [0] * len_array if len_array > 0: A = array[0] for i in range(1 , __UpperCAmelCase ): A = self.prefix_sum[i - 1] + array[i] def lowerCamelCase ( self :Dict , __UpperCamelCase :List[str] , __UpperCamelCase :Union[str, Any] ): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowerCamelCase ( self :Optional[Any] , __UpperCamelCase :Tuple ): A = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(__UpperCAmelCase ) return False if __name__ == "__main__": import doctest doctest.testmod()
292
"""simple docstring""" import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL UpperCamelCase : Union[str, Any] = version.parse(version.parse(torch.__version__).base_version) < version.parse("1.11") def A ( snake_case :str , snake_case :tuple , snake_case :Path , snake_case :Dict , snake_case :int , snake_case :List[str] , snake_case :Union[str, Any] , snake_case :Union[str, Any]=False , ) -> str: output_path.parent.mkdir(parents=snake_case , exist_ok=snake_case ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( snake_case , snake_case , f=output_path.as_posix() , input_names=snake_case , output_names=snake_case , dynamic_axes=snake_case , do_constant_folding=snake_case , use_external_data_format=snake_case , enable_onnx_checker=snake_case , opset_version=snake_case , ) else: export( snake_case , snake_case , f=output_path.as_posix() , input_names=snake_case , output_names=snake_case , dynamic_axes=snake_case , do_constant_folding=snake_case , opset_version=snake_case , ) @torch.no_grad() def A ( snake_case :str , snake_case :str , snake_case :int , snake_case :bool = False ) -> List[str]: __UpperCamelCase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): __UpperCamelCase = 'cuda' elif fpaa and not torch.cuda.is_available(): raise ValueError('`float16` model export is only supported on GPUs with CUDA' ) else: __UpperCamelCase = 'cpu' __UpperCamelCase = Path(snake_case ) # VAE DECODER __UpperCamelCase = AutoencoderKL.from_pretrained(model_path + '/vae' ) __UpperCamelCase = vae_decoder.config.latent_channels # forward only through the decoder part __UpperCamelCase = vae_decoder.decode onnx_export( snake_case , model_args=( torch.randn(1 , snake_case , 2_5 , 2_5 ).to(device=snake_case , dtype=snake_case ), False, ) , output_path=output_path / 'vae_decoder' / 'model.onnx' , ordered_input_names=['latent_sample', 'return_dict'] , output_names=['sample'] , dynamic_axes={ 'latent_sample': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, } , opset=snake_case , ) del vae_decoder if __name__ == "__main__": UpperCamelCase : Dict = argparse.ArgumentParser() parser.add_argument( "--model_path", type=str, required=True, help="Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).", ) parser.add_argument("--output_path", type=str, required=True, help="Path to the output model.") parser.add_argument( "--opset", default=1_4, type=int, help="The version of the ONNX operator set to use.", ) parser.add_argument("--fp16", action="store_true", default=False, help="Export the models in `float16` mode") UpperCamelCase : List[Any] = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("SD: Done: ONNX")
316
0
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : list[tuple[float, float]] ) -> Optional[Any]: '''simple docstring''' A: str = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. A: Dict = len(snake_case_ ) - 1 def _snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : float ) -> str: '''simple docstring''' assert 0 <= t <= 1, "Time t must be between 0 and 1." A: list[float] = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , snake_case_ ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(snake_case_ ) , 5 ) == 1 return output_values def _snake_case ( self : Dict , SCREAMING_SNAKE_CASE_ : float ) -> str: '''simple docstring''' assert 0 <= t <= 1, "Time t must be between 0 and 1." A: Any = self.basis_function(snake_case_ ) A: Dict = 0.0 A: Optional[Any] = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def _snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : float = 0.01 ) -> List[str]: '''simple docstring''' from matplotlib import pyplot as plt # type: ignore A: list[float] = [] # x coordinates of points to plot A: list[float] = [] # y coordinates of points to plot A: Tuple = 0.0 while t <= 1: A: str = self.bezier_curve_function(snake_case_ ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size A: List[str] = [i[0] for i in self.list_of_points] A: str = [i[1] for i in self.list_of_points] plt.plot( snake_case_ , snake_case_ , color='''blue''' , label='''Curve of Degree ''' + str(self.degree ) , ) plt.scatter(snake_case_ , snake_case_ , color='''red''' , label='''Control Points''' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
371
'''simple docstring''' import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowerCAmelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase_ : Optional[Any] = DebertaTokenizer UpperCamelCase_ : List[str] = True UpperCamelCase_ : int = DebertaTokenizerFast def _snake_case ( self : Optional[int] ) -> Dict: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A: Optional[int] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''[UNK]''', ] A: int = dict(zip(SCREAMING_SNAKE_CASE_ , range(len(SCREAMING_SNAKE_CASE_ ) ) ) ) A: Union[str, Any] = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] A: Union[str, Any] = {'''unk_token''': '''[UNK]'''} A: List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) A: str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(SCREAMING_SNAKE_CASE_ ) ) def _snake_case ( self : int , **SCREAMING_SNAKE_CASE_ : List[str] ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : Dict ) -> Union[str, Any]: '''simple docstring''' A: Optional[int] = '''lower newer''' A: str = '''lower newer''' return input_text, output_text def _snake_case ( self : Union[str, Any] ) -> Dict: '''simple docstring''' A: str = self.get_tokenizer() A: Any = '''lower newer''' A: Dict = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] A: int = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) A: List[Any] = tokens + [tokenizer.unk_token] A: int = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : List[Any] ) -> Any: '''simple docstring''' A: str = self.get_tokenizer() A: List[str] = tokenizer('''Hello''' , '''World''' ) A: Union[str, Any] = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd['''token_type_ids'''] , SCREAMING_SNAKE_CASE_ ) @slow def _snake_case ( self : Tuple ) -> Optional[int]: '''simple docstring''' A: Union[str, Any] = self.tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A: Any = tokenizer.encode('''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) A: Union[str, Any] = tokenizer.encode('''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) A: Dict = tokenizer.encode( '''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE_ , add_prefix_space=SCREAMING_SNAKE_CASE_ ) A: Dict = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE_ , add_prefix_space=SCREAMING_SNAKE_CASE_ ) A: List[Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ ) A: int = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def _snake_case ( self : Tuple ) -> Dict: '''simple docstring''' A: int = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: A: List[Any] = tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A: Dict = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] A: Dict = tokenizer(SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ ) A: Any = [tokenizer.decode(SCREAMING_SNAKE_CASE_ , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) for seq in encoding['''input_ids''']] # fmt: off A: Any = { '''input_ids''': [ [1, 21_18, 1_11_26, 5_65, 35, 83, 2_51_91, 1_63, 1_88_54, 13, 1_21_56, 12, 1_61_01, 2_53_76, 1_38_07, 9, 2_22_05, 2_78_93, 16_35, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 21_18, 1_11_26, 5_65, 2_45_36, 80, 4_37_97, 48_78, 73_73, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1_33, 78, 65, 16, 10, 37_24, 15_38, 3_31_83, 1_13_03, 4_37_97, 19_38, 4, 8_70, 2_41_65, 2_91_05, 5, 7_39, 3_26_44, 3_31_83, 1_13_03, 3_61_73, 88, 80, 6_50, 78_21, 4_59_40, 6, 52, 25_59, 5, 18_36, 9, 5, 73_97, 1_31_71, 31, 5, 18_36, 9, 3_26_44, 3_31_83, 1_13_03, 4, 2] ], '''token_type_ids''': [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], '''attention_mask''': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on A: Optional[int] = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] self.assertDictEqual(encoding.data , SCREAMING_SNAKE_CASE_ ) for expected, decoded in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
334
0
"""simple docstring""" import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging __a = ["bart.large", "bart.large.mnli", "bart.large.cnn", "bart_xsum/model.pt"] __a = {"bart.large": BartModel, "bart.large.mnli": BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse("0.9.0"): raise Exception("requires fairseq >= 0.9.0") logging.set_verbosity_info() __a = logging.get_logger(__name__) __a = " Hello world! cécé herlolip" __a = [ ("model.classification_heads.mnli.dense.weight", "classification_head.dense.weight"), ("model.classification_heads.mnli.dense.bias", "classification_head.dense.bias"), ("model.classification_heads.mnli.out_proj.weight", "classification_head.out_proj.weight"), ("model.classification_heads.mnli.out_proj.bias", "classification_head.out_proj.bias"), ] def A_ ( _lowercase ): '''simple docstring''' snake_case_ :List[Any] = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(_lowercase, _lowercase ) def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Optional[int] = dct.pop(_lowercase ) snake_case_ :Optional[Any] = val def A_ ( _lowercase ): '''simple docstring''' snake_case_ :Tuple = torch.load(_lowercase, map_location="""cpu""" ) snake_case_ :Union[str, Any] = torch.hub.load("""pytorch/fairseq""", """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def A_ ( _lowercase ): '''simple docstring''' snake_case_, snake_case_ :Any = emb.weight.shape snake_case_ :Dict = nn.Linear(_lowercase, _lowercase, bias=_lowercase ) snake_case_ :int = emb.weight.data return lin_layer @torch.no_grad() def A_ ( _lowercase, _lowercase, _lowercase=None ): '''simple docstring''' if not os.path.exists(_lowercase ): snake_case_ :str = torch.hub.load("""pytorch/fairseq""", _lowercase ).eval() else: snake_case_ :List[Any] = load_xsum_checkpoint(_lowercase ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: snake_case_ :Union[str, Any] = checkpoint_path.replace(""".""", """-""" ) snake_case_ :List[Any] = BartConfig.from_pretrained(_lowercase ) snake_case_ :Tuple = bart.encode(_lowercase ).unsqueeze(0 ) snake_case_ :Any = BartTokenizer.from_pretrained(_lowercase ).encode(_lowercase, return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(_lowercase, _lowercase ).all(): raise ValueError( f"""converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}""" ) if checkpoint_path == "bart.large.mnli": snake_case_ :int = bart.state_dict() remove_ignore_keys_(_lowercase ) snake_case_ :int = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(_lowercase, _lowercase, _lowercase ) snake_case_ :Optional[int] = BartForSequenceClassification(_lowercase ).eval() model.load_state_dict(_lowercase ) snake_case_ :int = bart.predict("""mnli""", _lowercase, return_logits=_lowercase ) snake_case_ :str = model(_lowercase )[0] # logits else: # no classification heads to worry about snake_case_ :Dict = bart.model.state_dict() remove_ignore_keys_(_lowercase ) snake_case_ :int = state_dict["""decoder.embed_tokens.weight"""] snake_case_ :str = bart.extract_features(_lowercase ) if hf_checkpoint_name == "facebook/bart-large": snake_case_ :Optional[Any] = BartModel(_lowercase ).eval() model.load_state_dict(_lowercase ) snake_case_ :Optional[Any] = model(_lowercase ).model[0] else: snake_case_ :List[Any] = BartForConditionalGeneration(_lowercase ).eval() # an existing summarization ckpt model.model.load_state_dict(_lowercase ) if hasattr(_lowercase, """lm_head""" ): snake_case_ :Tuple = make_linear_from_emb(model.model.shared ) snake_case_ :Optional[int] = model.model(_lowercase )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f"""`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}""" ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(_lowercase ).mkdir(exist_ok=_lowercase ) model.save_pretrained(_lowercase ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "fairseq_path", type=str, help="bart.large, bart.large.cnn or a path to a model.pt on local filesystem." ) parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--hf_config", default=None, type=str, help="Which huggingface architecture to use: bart-large-xsum" ) __a = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
66
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: snake_case__ : str = None snake_case__ : Optional[Any] = logging.get_logger(__name__) snake_case__ : Optional[int] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} snake_case__ : Dict = { '''vocab_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/spiece.model''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json''', }, } snake_case__ : Any = { '''google/fnet-base''': 512, '''google/fnet-large''': 512, } snake_case__ : Dict = '''▁''' class snake_case_( a__ ): __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ['''input_ids''', '''token_type_ids'''] __UpperCamelCase = FNetTokenizer def __init__( self : Union[str, Any] , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Any=False , UpperCamelCase_ : Any=True , UpperCamelCase_ : Dict=True , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="[SEP]" , UpperCamelCase_ : List[Any]="<pad>" , UpperCamelCase_ : Union[str, Any]="[CLS]" , UpperCamelCase_ : int="[MASK]" , **UpperCamelCase_ : Optional[Any] , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCAmelCase : int = ( AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ , normalized=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else mask_token ) super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , remove_space=UpperCamelCase_ , keep_accents=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , **UpperCamelCase_ , ) lowerCAmelCase : Optional[int] = do_lower_case lowerCAmelCase : str = remove_space lowerCAmelCase : Any = keep_accents lowerCAmelCase : int = vocab_file lowerCAmelCase : List[str] = False if not self.vocab_file else True def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None ): lowerCAmelCase : Optional[int] = [self.sep_token_id] lowerCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None ): lowerCAmelCase : List[str] = [self.sep_token_id] lowerCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase : str = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ): copyfile(self.vocab_file , UpperCamelCase_ ) return (out_vocab_file,)
60
0
"""simple docstring""" import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class lowerCAmelCase ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Dict = 1 @register_to_config def __init__( self , lowerCAmelCase__=2_000 , lowerCAmelCase__=0.1 , lowerCAmelCase__=20 , lowerCAmelCase__=1e-3 ) -> Dict: SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[Any]: SCREAMING_SNAKE_CASE = torch.linspace(1 , self.config.sampling_eps , lowercase__ , device=lowercase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ) -> Tuple: if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score SCREAMING_SNAKE_CASE = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) SCREAMING_SNAKE_CASE = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) SCREAMING_SNAKE_CASE = std.flatten() while len(std.shape ) < len(score.shape ): SCREAMING_SNAKE_CASE = std.unsqueeze(-1 ) SCREAMING_SNAKE_CASE = -score / std # compute SCREAMING_SNAKE_CASE = -1.0 / len(self.timesteps ) SCREAMING_SNAKE_CASE = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) SCREAMING_SNAKE_CASE = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): SCREAMING_SNAKE_CASE = beta_t.unsqueeze(-1 ) SCREAMING_SNAKE_CASE = -0.5 * beta_t * x SCREAMING_SNAKE_CASE = torch.sqrt(lowercase__ ) SCREAMING_SNAKE_CASE = drift - diffusion**2 * score SCREAMING_SNAKE_CASE = x + drift * dt # add noise SCREAMING_SNAKE_CASE = randn_tensor(x.shape , layout=x.layout , generator=lowercase__ , device=x.device , dtype=x.dtype ) SCREAMING_SNAKE_CASE = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ) -> Tuple: return self.config.num_train_timesteps
357
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __UpperCamelCase = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
38
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {"""vocab_file""": """spiece.model"""} lowerCAmelCase = { """vocab_file""": { """TsinghuaAI/CPM-Generate""": """https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model""", } } class A_ ( A__ ): """simple docstring""" def __init__( self :Union[str, Any] , lowerCamelCase_ :Any , lowerCamelCase_ :str=False , lowerCamelCase_ :Any=True , lowerCamelCase_ :int=False , lowerCamelCase_ :Optional[Any]="<s>" , lowerCamelCase_ :List[str]="</s>" , lowerCamelCase_ :List[Any]="<unk>" , lowerCamelCase_ :Union[str, Any]="<sep>" , lowerCamelCase_ :Union[str, Any]="<pad>" , lowerCamelCase_ :List[Any]="<cls>" , lowerCamelCase_ :List[str]="<mask>" , lowerCamelCase_ :Optional[int]=["<eop>", "<eod>"] , lowerCamelCase_ :Optional[Dict[str, Any]] = None , **lowerCamelCase_ :Tuple , ): """simple docstring""" lowerCamelCase__ : List[Any] =AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else mask_token lowerCamelCase__ : List[str] ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , additional_special_tokens=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) lowerCamelCase__ : List[str] =3 lowerCamelCase__ : Union[str, Any] =do_lower_case lowerCamelCase__ : str =remove_space lowerCamelCase__ : Optional[Any] =keep_accents lowerCamelCase__ : List[str] =vocab_file lowerCamelCase__ : int =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase_ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( 'You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ' 'See https://pypi.org/project/jieba/ for installation.' ) lowerCamelCase__ : List[str] =jieba lowerCamelCase__ : Tuple =str.maketrans(' \n' , '\u2582\u2583' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def UpperCAmelCase__ ( self :List[str] ): """simple docstring""" return len(self.sp_model ) def UpperCAmelCase__ ( self :List[Any] ): """simple docstring""" lowerCamelCase__ : Dict ={self.convert_ids_to_tokens(lowerCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self :Optional[Any] ): """simple docstring""" lowerCamelCase__ : Tuple =self.__dict__.copy() lowerCamelCase__ : Union[str, Any] =None return state def __setstate__( self :Optional[Any] , lowerCamelCase_ :Union[str, Any] ): """simple docstring""" lowerCamelCase__ : Optional[Any] =d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCamelCase__ : str ={} lowerCamelCase__ : Optional[Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase__ ( self :Union[str, Any] , lowerCamelCase_ :Any ): """simple docstring""" if self.remove_space: lowerCamelCase__ : Optional[Any] =' '.join(inputs.strip().split() ) else: lowerCamelCase__ : List[str] =inputs lowerCamelCase__ : Union[str, Any] =outputs.replace('``' , '"' ).replace('\'\'' , '"' ) if not self.keep_accents: lowerCamelCase__ : Union[str, Any] =unicodedata.normalize('NFKD' , lowerCamelCase_ ) lowerCamelCase__ : Dict =''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase_ )] ) if self.do_lower_case: lowerCamelCase__ : Dict =outputs.lower() return outputs def UpperCAmelCase__ ( self :Optional[Any] , lowerCamelCase_ :str ): """simple docstring""" lowerCamelCase__ : Tuple =self.preprocess_text(lowerCamelCase_ ) lowerCamelCase__ : str =self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_ ) lowerCamelCase__ : str =[] for piece in pieces: if len(lowerCamelCase_ ) > 1 and piece[-1] == str(',' ) and piece[-2].isdigit(): lowerCamelCase__ : Optional[int] =self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase_ , '' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase__ : Any =cur_pieces[1:] else: lowerCamelCase__ : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(lowerCamelCase_ ) else: new_pieces.append(lowerCamelCase_ ) return new_pieces def UpperCAmelCase__ ( self :List[Any] , lowerCamelCase_ :int ): """simple docstring""" return self.sp_model.PieceToId(lowerCamelCase_ ) def UpperCAmelCase__ ( self :int , lowerCamelCase_ :str ): """simple docstring""" return self.sp_model.IdToPiece(lowerCamelCase_ ) def UpperCAmelCase__ ( self :Dict , lowerCamelCase_ :List[str] ): """simple docstring""" lowerCamelCase__ : str =''.join(lowerCamelCase_ ).replace(lowerCamelCase_ , ' ' ).strip() return out_string def UpperCAmelCase__ ( self :Dict , lowerCamelCase_ :List[int] , lowerCamelCase_ :Optional[List[int]] = None ): """simple docstring""" lowerCamelCase__ : List[str] =[self.sep_token_id] lowerCamelCase__ : int =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase__ ( self :Union[str, Any] , lowerCamelCase_ :List[int] , lowerCamelCase_ :Optional[List[int]] = None , lowerCamelCase_ :bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is not None: return ([0] * len(lowerCamelCase_ )) + [1] + ([0] * len(lowerCamelCase_ )) + [1, 1] return ([0] * len(lowerCamelCase_ )) + [1, 1] def UpperCAmelCase__ ( self :int , lowerCamelCase_ :List[int] , lowerCamelCase_ :Optional[List[int]] = None ): """simple docstring""" lowerCamelCase__ : Optional[Any] =[self.sep_token_id] lowerCamelCase__ : str =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCAmelCase__ ( self :Tuple , lowerCamelCase_ :str , lowerCamelCase_ :Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : Optional[Any] =os.path.join( lowerCamelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase_ , 'wb' ) as fi: lowerCamelCase__ : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_ ) return (out_vocab_file,) def UpperCAmelCase__ ( self :int , *lowerCamelCase_ :List[Any] , **lowerCamelCase_ :Dict ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =super()._decode(*lowerCamelCase_ , **lowerCamelCase_ ) lowerCamelCase__ : Tuple =text.replace(' ' , '' ).replace('\u2582' , ' ' ).replace('\u2583' , '\n' ) return text
126
"""simple docstring""" import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class A_ : """simple docstring""" @staticmethod def UpperCAmelCase__ ( *lowerCamelCase_ :Optional[int] , **lowerCamelCase_ :Optional[Any] ): """simple docstring""" pass def lowerCAmelCase_ ( snake_case_ : List[str] ) ->str: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. lowerCAmelCase = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class A_ ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Optional[Any] , lowerCamelCase_ :Union[str, Any] , lowerCamelCase_ :Tuple , lowerCamelCase_ :int ): """simple docstring""" lowerCamelCase__ : Optional[int] =pipeline( 'document-question-answering' , model=lowerCamelCase_ , tokenizer=lowerCamelCase_ , image_processor=lowerCamelCase_ ) lowerCamelCase__ : Tuple =INVOICE_URL lowerCamelCase__ : Optional[Any] =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) lowerCamelCase__ : Optional[Any] ='What is the placebo?' lowerCamelCase__ : List[str] =[ { 'image': load_image(lowerCamelCase_ ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def UpperCAmelCase__ ( self :int , lowerCamelCase_ :Tuple , lowerCamelCase_ :Tuple ): """simple docstring""" lowerCamelCase__ : List[str] =dqa_pipeline(lowerCamelCase_ , top_k=2 ) self.assertEqual( lowerCamelCase_ , [ [ {'score': ANY(lowerCamelCase_ ), 'answer': ANY(lowerCamelCase_ ), 'start': ANY(lowerCamelCase_ ), 'end': ANY(lowerCamelCase_ )}, {'score': ANY(lowerCamelCase_ ), 'answer': ANY(lowerCamelCase_ ), 'start': ANY(lowerCamelCase_ ), 'end': ANY(lowerCamelCase_ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : str =pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) lowerCamelCase__ : Any =INVOICE_URL lowerCamelCase__ : Union[str, Any] ='How many cats are there?' lowerCamelCase__ : List[Any] =[ {'score': 0.00_01, 'answer': 'oy 2312/2019', 'start': 38, 'end': 39}, {'score': 0.00_01, 'answer': 'oy 2312/2019 DUE', 'start': 38, 'end': 40}, ] lowerCamelCase__ : Dict =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , lowerCamelCase_ ) lowerCamelCase__ : int =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , lowerCamelCase_ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowerCamelCase__ : str ='./tests/fixtures/tests_samples/COCO/000000039769.png' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(lowerCamelCase_ , [] ) # We can optionnally pass directly the words and bounding boxes lowerCamelCase__ : str ='./tests/fixtures/tests_samples/COCO/000000039769.png' lowerCamelCase__ : Optional[Any] =[] lowerCamelCase__ : Tuple =[] lowerCamelCase__ : Tuple =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , words=lowerCamelCase_ , boxes=lowerCamelCase_ , top_k=2 ) self.assertEqual(lowerCamelCase_ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : int =pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) lowerCamelCase__ : Dict =INVOICE_URL lowerCamelCase__ : int ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[str] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :List[str] ): """simple docstring""" lowerCamelCase__ : int =pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=50 , ) lowerCamelCase__ : Tuple =INVOICE_URL lowerCamelCase__ : Any ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[Any] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[Any] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Dict ): """simple docstring""" lowerCamelCase__ : int =AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCamelCase_ , revision='3dc6de3' , ) lowerCamelCase__ : int =INVOICE_URL lowerCamelCase__ : Tuple ='What is the invoice number?' lowerCamelCase__ : Optional[Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowerCamelCase__ : Optional[Any] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] ] * 2 , ) lowerCamelCase__ : Tuple =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) # This model should also work if `image` is set to None lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Optional[Any] ): """simple docstring""" lowerCamelCase__ : Optional[Any] =AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCamelCase_ ) lowerCamelCase__ : Any =pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCamelCase_ , revision='3dc6de3' , max_seq_len=50 , ) lowerCamelCase__ : Dict =INVOICE_URL lowerCamelCase__ : Optional[Any] ='What is the invoice number?' lowerCamelCase__ : Tuple =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : Tuple =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) lowerCamelCase__ : str =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) # This model should also work if `image` is set to None lowerCamelCase__ : Union[str, Any] =dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) @slow @require_torch def UpperCAmelCase__ ( self :str ): """simple docstring""" lowerCamelCase__ : List[Any] =pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) lowerCamelCase__ : Union[str, Any] =INVOICE_URL lowerCamelCase__ : Union[str, Any] ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def UpperCAmelCase__ ( self :str ): """simple docstring""" pass
126
1
'''simple docstring''' import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class A_ ( _snake_case ): '''simple docstring''' UpperCAmelCase_ : str = (DPMSolverSDEScheduler,) UpperCAmelCase_ : int = 10 def UpperCAmelCase_ ( self : Tuple , **lowercase_ : Tuple ) -> List[Any]: UpperCAmelCase : str = { 'num_train_timesteps': 1_100, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'noise_sampler_seed': 0, } config.update(**lowerCamelCase_ ) return config def UpperCAmelCase_ ( self : Optional[Any] ) -> str: for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=lowerCamelCase_ ) def UpperCAmelCase_ ( self : Any ) -> int: for beta_start, beta_end in zip([0.0_0001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=lowerCamelCase_ , beta_end=lowerCamelCase_ ) def UpperCAmelCase_ ( self : List[str] ) -> List[Any]: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCamelCase_ ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase_ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: UpperCAmelCase : List[str] = self.scheduler_classes[0] UpperCAmelCase : str = self.get_scheduler_config() UpperCAmelCase : Dict = scheduler_class(**lowerCamelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) UpperCAmelCase : List[str] = self.dummy_model() UpperCAmelCase : Optional[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma UpperCAmelCase : int = sample.to(lowerCamelCase_ ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase : Union[str, Any] = scheduler.scale_model_input(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Tuple = model(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : str = scheduler.step(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Dict = output.prev_sample UpperCAmelCase : Optional[Any] = torch.sum(torch.abs(lowerCamelCase_ ) ) UpperCAmelCase : str = torch.mean(torch.abs(lowerCamelCase_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.47_8210_4492_1875 ) < 1E-2 assert abs(result_mean.item() - 0.2178_7059_6456_5277 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_3521_1181_6406 ) < 1E-2 assert abs(result_mean.item() - 0.2_2342_9068_9229_9652 ) < 1E-3 else: assert abs(result_sum.item() - 162.52_3834_2285_1562 ) < 1E-2 assert abs(result_mean.item() - 0.211_6195_7085_1326 ) < 1E-3 def UpperCAmelCase_ ( self : Any ) -> int: UpperCAmelCase : int = self.scheduler_classes[0] UpperCAmelCase : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) UpperCAmelCase : List[Any] = scheduler_class(**lowerCamelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) UpperCAmelCase : Dict = self.dummy_model() UpperCAmelCase : int = self.dummy_sample_deter * scheduler.init_noise_sigma UpperCAmelCase : Union[str, Any] = sample.to(lowerCamelCase_ ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase : Union[str, Any] = scheduler.scale_model_input(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Tuple = model(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : int = scheduler.step(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Any = output.prev_sample UpperCAmelCase : str = torch.sum(torch.abs(lowerCamelCase_ ) ) UpperCAmelCase : Any = torch.mean(torch.abs(lowerCamelCase_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 124.77_1492_0043_9453 ) < 1E-2 assert abs(result_mean.item() - 0.1_6226_2890_1481_6284 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 128.1_6633_6059_5703 ) < 1E-2 assert abs(result_mean.item() - 0.1_6688_3260_0116_7297 ) < 1E-3 else: assert abs(result_sum.item() - 119.8_4875_4882_8125 ) < 1E-2 assert abs(result_mean.item() - 0.1560_5306_6253_6621 ) < 1E-3 def UpperCAmelCase_ ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.scheduler_classes[0] UpperCAmelCase : Optional[Any] = self.get_scheduler_config() UpperCAmelCase : List[Any] = scheduler_class(**lowerCamelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=lowerCamelCase_ ) UpperCAmelCase : Optional[int] = self.dummy_model() UpperCAmelCase : List[Any] = self.dummy_sample_deter.to(lowerCamelCase_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: UpperCAmelCase : int = scheduler.scale_model_input(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Optional[Any] = model(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Tuple = scheduler.step(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Optional[int] = output.prev_sample UpperCAmelCase : Optional[Any] = torch.sum(torch.abs(lowerCamelCase_ ) ) UpperCAmelCase : Dict = torch.mean(torch.abs(lowerCamelCase_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.46_9573_9746_0938 ) < 1E-2 assert abs(result_mean.item() - 0.2_1805_9346_0798_2635 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_3536_3769_5312 ) < 1E-2 assert abs(result_mean.item() - 0.2_2342_9083_8241_5771 ) < 1E-3 else: assert abs(result_sum.item() - 162.52_3834_2285_1562 ) < 1E-2 assert abs(result_mean.item() - 0.211_6195_7085_1326 ) < 1E-3 def UpperCAmelCase_ ( self : int ) -> List[Any]: UpperCAmelCase : Dict = self.scheduler_classes[0] UpperCAmelCase : Dict = self.get_scheduler_config() UpperCAmelCase : Union[str, Any] = scheduler_class(**lowerCamelCase_ , use_karras_sigmas=lowerCamelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=lowerCamelCase_ ) UpperCAmelCase : Dict = self.dummy_model() UpperCAmelCase : Optional[Any] = self.dummy_sample_deter.to(lowerCamelCase_ ) * scheduler.init_noise_sigma UpperCAmelCase : Optional[Any] = sample.to(lowerCamelCase_ ) for t in scheduler.timesteps: UpperCAmelCase : str = scheduler.scale_model_input(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Optional[int] = model(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : List[str] = scheduler.step(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase : Dict = output.prev_sample UpperCAmelCase : Union[str, Any] = torch.sum(torch.abs(lowerCamelCase_ ) ) UpperCAmelCase : Any = torch.mean(torch.abs(lowerCamelCase_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 176.66_9741_3574_2188 ) < 1E-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1E-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 177.63_6535_6445_3125 ) < 1E-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1E-2 else: assert abs(result_sum.item() - 170.3_1352_2338_8672 ) < 1E-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1E-2
355
'''simple docstring''' import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller lowercase__ = 3 def UpperCamelCase( UpperCAmelCase_ ): print('Generating primitive root of p' ) while True: UpperCAmelCase : Union[str, Any] = random.randrange(3 , UpperCAmelCase_ ) if pow(UpperCAmelCase_ , 2 , UpperCAmelCase_ ) == 1: continue if pow(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) == 1: continue return g def UpperCamelCase( UpperCAmelCase_ ): print('Generating prime p...' ) UpperCAmelCase : str = rabin_miller.generate_large_prime(UpperCAmelCase_ ) # select large prime number. UpperCAmelCase : List[str] = primitive_root(UpperCAmelCase_ ) # one primitive root on modulo p. UpperCAmelCase : List[Any] = random.randrange(3 , UpperCAmelCase_ ) # private_key -> have to be greater than 2 for safety. UpperCAmelCase : List[Any] = cryptomath.find_mod_inverse(pow(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) UpperCAmelCase : Tuple = (key_size, e_a, e_a, p) UpperCAmelCase : Optional[int] = (key_size, d) return public_key, private_key def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): if os.path.exists(F"""{name}_pubkey.txt""" ) or os.path.exists(F"""{name}_privkey.txt""" ): print('\nWARNING:' ) print( F"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" 'Use a different name or delete these files and re-run this program.' ) sys.exit() UpperCAmelCase , UpperCAmelCase : Dict = generate_key(UpperCAmelCase_ ) print(F"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(F"""{name}_pubkey.txt""" , 'w' ) as fo: fo.write(F"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(F"""Writing private key to file {name}_privkey.txt...""" ) with open(F"""{name}_privkey.txt""" , 'w' ) as fo: fo.write(F"""{private_key[0]},{private_key[1]}""" ) def UpperCamelCase( ): print('Making key files...' ) make_key_files('elgamal' , 20_48 ) print('Key files generation successful' ) if __name__ == "__main__": main()
280
0
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
def UpperCAmelCase ( a_ = 6_0_0_8_5_1_4_7_5_1_4_3 ) -> str: """simple docstring""" try: __A = int(UpperCamelCase__ ) except (TypeError, ValueError): raise TypeError("Parameter n must be int or castable to int." ) if n <= 0: raise ValueError("Parameter n must be greater than or equal to one." ) __A = 2 __A = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 __A = i while n % i == 0: __A = n // i i += 1 return int(UpperCamelCase__ ) if __name__ == "__main__": print(f'''{solution() = }''')
352
def UpperCAmelCase ( ) -> list[list[int]]: """simple docstring""" return [list(range(1_0_0_0 - i , -1_0_0_0 - i , -1 ) ) for i in range(1_0_0_0 )] SCREAMING_SNAKE_CASE :List[str] = generate_large_matrix() SCREAMING_SNAKE_CASE :str = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def UpperCAmelCase ( a_ ) -> None: """simple docstring""" assert all(row == sorted(a_ , reverse=a_ ) for row in grid ) assert all(list(a_ ) == sorted(a_ , reverse=a_ ) for col in zip(*a_ ) ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" __A = 0 __A = len(a_ ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: __A = (left + right) // 2 __A = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: __A = mid + 1 else: __A = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" __A = 0 __A = len(grid[0] ) for i in range(len(a_ ) ): __A = find_negative_index(grid[i][:bound] ) total += bound return (len(a_ ) * len(grid[0] )) - total def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" __A = 0 for row in grid: for i, number in enumerate(a_ ): if number < 0: total += len(a_ ) - i break return total def UpperCAmelCase ( ) -> None: """simple docstring""" from timeit import timeit print("Running benchmarks" ) __A = ( "from __main__ import count_negatives_binary_search, " "count_negatives_brute_force, count_negatives_brute_force_with_break, grid" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): __A = timeit(F'''{func}(grid=grid)''' , setup=a_ , number=5_0_0 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
124
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case__ : Any = logging.get_logger(__name__) def _snake_case ( _snake_case : List[str] ): lowerCAmelCase : Union[str, Any] = '''huggingface/label-files''' lowerCAmelCase : Dict = '''imagenet-1k-id2label.json''' lowerCAmelCase : List[Any] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase : str = {int(_snake_case ): v for k, v in idalabel.items()} lowerCAmelCase : Any = {v: k for k, v in idalabel.items()} lowerCAmelCase : List[str] = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCAmelCase : List[Any] = BitConfig( conv_layer=_snake_case , num_labels=1000 , idalabel=_snake_case , labelaid=_snake_case , ) return config def _snake_case ( _snake_case : Dict ): if "stem.conv" in name: lowerCAmelCase : Optional[int] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: lowerCAmelCase : List[Any] = name.replace('''blocks''' , '''layers''' ) if "head.fc" in name: lowerCAmelCase : Optional[int] = name.replace('''head.fc''' , '''classifier.1''' ) if name.startswith('''norm''' ): lowerCAmelCase : Union[str, Any] = '''bit.''' + name if "bit" not in name and "classifier" not in name: lowerCAmelCase : Any = '''bit.encoder.''' + name return name def _snake_case ( ): lowerCAmelCase : List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase : str = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def _snake_case ( _snake_case : List[Any] , _snake_case : str , _snake_case : List[str]=False ): lowerCAmelCase : Optional[Any] = get_config(_snake_case ) # load original model from timm lowerCAmelCase : List[Any] = create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model lowerCAmelCase : str = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCAmelCase : Dict = state_dict.pop(_snake_case ) lowerCAmelCase : Any = val.squeeze() if '''head''' in key else val # load HuggingFace model lowerCAmelCase : Optional[int] = BitForImageClassification(_snake_case ) model.eval() model.load_state_dict(_snake_case ) # create image processor lowerCAmelCase : Union[str, Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) lowerCAmelCase : int = transform.transforms lowerCAmelCase : int = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } lowerCAmelCase : Any = BitImageProcessor( do_resize=_snake_case , size={'''shortest_edge''': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCAmelCase : List[str] = prepare_img() lowerCAmelCase : Dict = transform(_snake_case ).unsqueeze(0 ) lowerCAmelCase : Tuple = processor(_snake_case , return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): lowerCAmelCase : Dict = model(_snake_case ) lowerCAmelCase : Optional[Any] = outputs.logits print('''Logits:''' , logits[0, :3] ) print('''Predicted class:''' , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCAmelCase : Dict = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) processor.save_pretrained(_snake_case ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": snake_case__ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''resnetv2_50x1_bitm''', type=str, help='''Name of the BiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model to the hub.''', ) snake_case__ : Any = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
60
"""simple docstring""" from __future__ import annotations from decimal import Decimal from numpy import array def _snake_case ( _snake_case : list[list[float]] ): lowerCAmelCase : str = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(_snake_case ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix lowerCAmelCase : int = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creates a copy of the matrix with swapped positions of the elements lowerCAmelCase : Optional[int] = [[0.0, 0.0], [0.0, 0.0]] lowerCAmelCase, lowerCAmelCase : List[Any] = matrix[1][1], matrix[0][0] lowerCAmelCase, lowerCAmelCase : Union[str, Any] = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(_snake_case ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(_snake_case ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule lowerCAmelCase : int = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creating cofactor matrix lowerCAmelCase : Dict = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] lowerCAmelCase : List[str] = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) lowerCAmelCase : Dict = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) lowerCAmelCase : str = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) lowerCAmelCase : Any = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) lowerCAmelCase : Any = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) lowerCAmelCase : Optional[int] = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) lowerCAmelCase : Optional[int] = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) lowerCAmelCase : Dict = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) lowerCAmelCase : List[Any] = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) lowerCAmelCase : str = array(_snake_case ) for i in range(3 ): for j in range(3 ): lowerCAmelCase : Optional[Any] = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix lowerCAmelCase : Tuple = array(_snake_case ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(_snake_case ) # Calculate the inverse of the matrix return [[float(d(_snake_case ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError('''Please provide a matrix of size 2x2 or 3x3.''' )
60
1
import functools def _A ( lowerCAmelCase_ : str , lowerCAmelCase_ : str ): """simple docstring""" lowerCAmelCase__ = len(__SCREAMING_SNAKE_CASE ) lowerCAmelCase__ = len(__SCREAMING_SNAKE_CASE ) @functools.cache def min_distance(lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa lowerCAmelCase__ = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , __SCREAMING_SNAKE_CASE ) , 1 + min_distance(__SCREAMING_SNAKE_CASE , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
370
from math import pi, sqrt def _A ( lowerCAmelCase_ : float ): """simple docstring""" if num <= 0: raise ValueError("math domain error" ) if num > 171.5: raise OverflowError("math range error" ) elif num - int(lowerCAmelCase_ ) not in (0, 0.5): raise NotImplementedError("num must be an integer or a half-integer" ) elif num == 0.5: return sqrt(lowerCAmelCase_ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def _A ( ): """simple docstring""" assert gamma(0.5 ) == sqrt(lowerCAmelCase_ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = 1.0 while num: UpperCamelCase = float(input('Gamma of: ')) print(F"""gamma({num}) = {gamma(num)}""") print('\nEnter 0 to exit...')
221
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging A_ = logging.get_logger(__name__) if is_vision_available(): import PIL class _snake_case ( _a ): _A : List[Any] = ['''pixel_values'''] def __init__( self : Optional[Any] ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : Dict[str, int] = None ,SCREAMING_SNAKE_CASE__ : PILImageResampling = PILImageResampling.BICUBIC ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : Dict[str, int] = None ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : Union[int, float] = 1 / 255 ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None ,SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None ,SCREAMING_SNAKE_CASE__ : bool = True ,**SCREAMING_SNAKE_CASE__ : str ,): super().__init__(**SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Union[str, Any] = size if size is not None else {"shortest_edge": 224} SCREAMING_SNAKE_CASE:Optional[int] = get_size_dict(SCREAMING_SNAKE_CASE__ ,default_to_square=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE:Optional[int] = get_size_dict(SCREAMING_SNAKE_CASE__ ,default_to_square=SCREAMING_SNAKE_CASE__ ,param_name="crop_size" ) SCREAMING_SNAKE_CASE:int = do_resize SCREAMING_SNAKE_CASE:Dict = size SCREAMING_SNAKE_CASE:Tuple = resample SCREAMING_SNAKE_CASE:List[str] = do_center_crop SCREAMING_SNAKE_CASE:Optional[Any] = crop_size SCREAMING_SNAKE_CASE:List[Any] = do_rescale SCREAMING_SNAKE_CASE:Any = rescale_factor SCREAMING_SNAKE_CASE:List[str] = do_normalize SCREAMING_SNAKE_CASE:Union[str, Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN SCREAMING_SNAKE_CASE:int = image_std if image_std is not None else OPENAI_CLIP_STD SCREAMING_SNAKE_CASE:List[Any] = do_convert_rgb def __UpperCamelCase ( self : Any ,SCREAMING_SNAKE_CASE__ : np.ndarray ,SCREAMING_SNAKE_CASE__ : Dict[str, int] ,SCREAMING_SNAKE_CASE__ : PILImageResampling = PILImageResampling.BICUBIC ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None ,**SCREAMING_SNAKE_CASE__ : Any ,): SCREAMING_SNAKE_CASE:List[Any] = get_size_dict(SCREAMING_SNAKE_CASE__ ,default_to_square=SCREAMING_SNAKE_CASE__ ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) SCREAMING_SNAKE_CASE:Any = get_resize_output_image_size(SCREAMING_SNAKE_CASE__ ,size=size["shortest_edge"] ,default_to_square=SCREAMING_SNAKE_CASE__ ) return resize(SCREAMING_SNAKE_CASE__ ,size=SCREAMING_SNAKE_CASE__ ,resample=SCREAMING_SNAKE_CASE__ ,data_format=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : Any ,SCREAMING_SNAKE_CASE__ : np.ndarray ,SCREAMING_SNAKE_CASE__ : Dict[str, int] ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None ,**SCREAMING_SNAKE_CASE__ : str ,): SCREAMING_SNAKE_CASE:Union[str, Any] = get_size_dict(SCREAMING_SNAKE_CASE__ ) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(SCREAMING_SNAKE_CASE__ ,size=(size["height"], size["width"]) ,data_format=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : str ,SCREAMING_SNAKE_CASE__ : np.ndarray ,SCREAMING_SNAKE_CASE__ : Union[int, float] ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None ,**SCREAMING_SNAKE_CASE__ : Tuple ,): return rescale(SCREAMING_SNAKE_CASE__ ,scale=SCREAMING_SNAKE_CASE__ ,data_format=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : Optional[Any] ,SCREAMING_SNAKE_CASE__ : np.ndarray ,SCREAMING_SNAKE_CASE__ : Union[float, List[float]] ,SCREAMING_SNAKE_CASE__ : Union[float, List[float]] ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None ,**SCREAMING_SNAKE_CASE__ : Union[str, Any] ,): return normalize(SCREAMING_SNAKE_CASE__ ,mean=SCREAMING_SNAKE_CASE__ ,std=SCREAMING_SNAKE_CASE__ ,data_format=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : List[str] ,SCREAMING_SNAKE_CASE__ : ImageInput ,SCREAMING_SNAKE_CASE__ : bool = None ,SCREAMING_SNAKE_CASE__ : Dict[str, int] = None ,SCREAMING_SNAKE_CASE__ : PILImageResampling = None ,SCREAMING_SNAKE_CASE__ : bool = None ,SCREAMING_SNAKE_CASE__ : int = None ,SCREAMING_SNAKE_CASE__ : bool = None ,SCREAMING_SNAKE_CASE__ : float = None ,SCREAMING_SNAKE_CASE__ : bool = None ,SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None ,SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None ,SCREAMING_SNAKE_CASE__ : bool = None ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, TensorType]] = None ,SCREAMING_SNAKE_CASE__ : Optional[ChannelDimension] = ChannelDimension.FIRST ,**SCREAMING_SNAKE_CASE__ : List[str] ,): SCREAMING_SNAKE_CASE:Union[str, Any] = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE:Tuple = size if size is not None else self.size SCREAMING_SNAKE_CASE:List[Any] = get_size_dict(SCREAMING_SNAKE_CASE__ ,param_name="size" ,default_to_square=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Any = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE:Optional[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE:Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE:int = get_size_dict(SCREAMING_SNAKE_CASE__ ,param_name="crop_size" ,default_to_square=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE:List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE:str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE:int = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE:Union[str, Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE:Tuple = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb SCREAMING_SNAKE_CASE:Any = make_list_of_images(SCREAMING_SNAKE_CASE__ ) if not valid_images(SCREAMING_SNAKE_CASE__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: SCREAMING_SNAKE_CASE:Optional[int] = [convert_to_rgb(SCREAMING_SNAKE_CASE__ ) for image in images] # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE:int = [to_numpy_array(SCREAMING_SNAKE_CASE__ ) for image in images] if do_resize: SCREAMING_SNAKE_CASE:int = [self.resize(image=SCREAMING_SNAKE_CASE__ ,size=SCREAMING_SNAKE_CASE__ ,resample=SCREAMING_SNAKE_CASE__ ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE:Optional[int] = [self.center_crop(image=SCREAMING_SNAKE_CASE__ ,size=SCREAMING_SNAKE_CASE__ ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE:str = [self.rescale(image=SCREAMING_SNAKE_CASE__ ,scale=SCREAMING_SNAKE_CASE__ ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE:Any = [self.normalize(image=SCREAMING_SNAKE_CASE__ ,mean=SCREAMING_SNAKE_CASE__ ,std=SCREAMING_SNAKE_CASE__ ) for image in images] SCREAMING_SNAKE_CASE:List[Any] = [to_channel_dimension_format(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) for image in images] SCREAMING_SNAKE_CASE:Optional[int] = {"pixel_values": images} return BatchFeature(data=SCREAMING_SNAKE_CASE__ ,tensor_type=SCREAMING_SNAKE_CASE__ )
139
'''simple docstring''' def A_ ( snake_case = 100 ): SCREAMING_SNAKE_CASE:Optional[Any] = set() SCREAMING_SNAKE_CASE:int = 0 SCREAMING_SNAKE_CASE:Optional[Any] = n + 1 # maximum limit for a in range(2 , snake_case ): for b in range(2 , snake_case ): SCREAMING_SNAKE_CASE:Tuple = a**b # calculates the current power collect_powers.add(snake_case ) # adds the result to the set return len(snake_case ) if __name__ == "__main__": print("Number of terms ", solution(int(str(input()).strip())))
139
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Optional[int] = logging.get_logger(__name__) UpperCAmelCase : Union[str, Any] = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class lowerCamelCase__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __a = """sew""" def __init__( self : Dict , UpperCamelCase : Dict=32 , UpperCamelCase : Optional[int]=768 , UpperCamelCase : Tuple=12 , UpperCamelCase : List[Any]=12 , UpperCamelCase : str=3_072 , UpperCamelCase : Optional[Any]=2 , UpperCamelCase : int="gelu" , UpperCamelCase : Any=0.1 , UpperCamelCase : Optional[int]=0.1 , UpperCamelCase : Dict=0.1 , UpperCamelCase : List[str]=0.0 , UpperCamelCase : Union[str, Any]=0.1 , UpperCamelCase : Optional[Any]=0.1 , UpperCamelCase : List[str]=0.02 , UpperCamelCase : List[str]=1e-5 , UpperCamelCase : List[str]="group" , UpperCamelCase : str="gelu" , UpperCamelCase : Dict=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , UpperCamelCase : Tuple=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , UpperCamelCase : List[Any]=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , UpperCamelCase : str=False , UpperCamelCase : Optional[int]=128 , UpperCamelCase : Any=16 , UpperCamelCase : List[str]=True , UpperCamelCase : int=0.05 , UpperCamelCase : Tuple=10 , UpperCamelCase : Optional[int]=2 , UpperCamelCase : Tuple=0.0 , UpperCamelCase : int=10 , UpperCamelCase : Optional[Any]=0 , UpperCamelCase : List[Any]="mean" , UpperCamelCase : int=False , UpperCamelCase : Optional[Any]=False , UpperCamelCase : Tuple=256 , UpperCamelCase : Tuple=0 , UpperCamelCase : Dict=1 , UpperCamelCase : Optional[Any]=2 , **UpperCamelCase : List[str] , ): '''simple docstring''' super().__init__(**UpperCamelCase , pad_token_id=UpperCamelCase , bos_token_id=UpperCamelCase , eos_token_id=UpperCamelCase ) __UpperCAmelCase : List[str] = hidden_size __UpperCAmelCase : str = feat_extract_norm __UpperCAmelCase : Dict = feat_extract_activation __UpperCAmelCase : Union[str, Any] = list(UpperCamelCase ) __UpperCAmelCase : List[str] = list(UpperCamelCase ) __UpperCAmelCase : str = list(UpperCamelCase ) __UpperCAmelCase : Optional[Any] = conv_bias __UpperCAmelCase : Optional[int] = num_conv_pos_embeddings __UpperCAmelCase : List[Any] = num_conv_pos_embedding_groups __UpperCAmelCase : List[str] = len(self.conv_dim ) __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Optional[int] = intermediate_size __UpperCAmelCase : int = squeeze_factor __UpperCAmelCase : Dict = hidden_act __UpperCAmelCase : List[Any] = num_attention_heads __UpperCAmelCase : int = hidden_dropout __UpperCAmelCase : Optional[Any] = attention_dropout __UpperCAmelCase : Tuple = activation_dropout __UpperCAmelCase : Tuple = feat_proj_dropout __UpperCAmelCase : List[Any] = final_dropout __UpperCAmelCase : Dict = layerdrop __UpperCAmelCase : List[str] = layer_norm_eps __UpperCAmelCase : Optional[Any] = initializer_range __UpperCAmelCase : Tuple = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" f'''but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)''' f'''= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCAmelCase : Optional[Any] = apply_spec_augment __UpperCAmelCase : Any = mask_time_prob __UpperCAmelCase : Tuple = mask_time_length __UpperCAmelCase : Tuple = mask_time_min_masks __UpperCAmelCase : Dict = mask_feature_prob __UpperCAmelCase : Tuple = mask_feature_length __UpperCAmelCase : Tuple = mask_feature_min_masks # ctc loss __UpperCAmelCase : List[Any] = ctc_loss_reduction __UpperCAmelCase : Optional[Any] = ctc_zero_infinity # sequence classification __UpperCAmelCase : Dict = use_weighted_layer_sum __UpperCAmelCase : str = classifier_proj_size @property def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
351
"""simple docstring""" from __future__ import annotations def lowerCamelCase ( _UpperCamelCase : list[float] , _UpperCamelCase : list[float] ) -> float: '''simple docstring''' __UpperCAmelCase : Tuple = sorted(numsa + numsa ) __UpperCAmelCase ,__UpperCAmelCase : Dict = divmod(len(_UpperCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase : List[Any] = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase : Optional[int] = [float(x) for x in input('Enter the elements of second array: ').split()] print(F"The median of two arrays is: {median_of_two_arrays(array_a, array_a)}")
320
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class _lowerCAmelCase : # setable values __SCREAMING_SNAKE_CASE : Optional[int] = None __SCREAMING_SNAKE_CASE : Optional[jnp.ndarray] = None __SCREAMING_SNAKE_CASE : Optional[jnp.ndarray] = None # sigma(t_i) @classmethod def _a (cls ): return cls() @dataclass class _lowerCAmelCase ( __UpperCAmelCase ): __SCREAMING_SNAKE_CASE : jnp.ndarray __SCREAMING_SNAKE_CASE : jnp.ndarray __SCREAMING_SNAKE_CASE : KarrasVeSchedulerState class _lowerCAmelCase ( __UpperCAmelCase , __UpperCAmelCase ): @property def _a (self ): return True @register_to_config def __init__(self , lowercase = 0.02 , lowercase = 100 , lowercase = 1.0_07 , lowercase = 80 , lowercase = 0.05 , lowercase = 50 , ): pass def _a (self ): return KarrasVeSchedulerState.create() def _a (self , lowercase , lowercase , lowercase = () ): A_ : Any = jnp.arange(0 , lowercase )[::-1].copy() A_ : Any = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=lowercase , schedule=jnp.array(lowercase , dtype=jnp.floataa ) , timesteps=lowercase , ) def _a (self , lowercase , lowercase , lowercase , lowercase , ): if self.config.s_min <= sigma <= self.config.s_max: A_ : Any = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: A_ : Tuple = 0 # sample eps ~ N(0, S_noise^2 * I) A_ : Dict = random.split(lowercase , num=1 ) A_ : Tuple = self.config.s_noise * random.normal(key=lowercase , shape=sample.shape ) A_ : str = sigma + gamma * sigma A_ : str = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = True , ): A_ : Union[str, Any] = sample_hat + sigma_hat * model_output A_ : List[str] = (sample_hat - pred_original_sample) / sigma_hat A_ : int = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowercase , derivative=lowercase , state=lowercase ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = True , ): A_ : Dict = sample_prev + sigma_prev * model_output A_ : Optional[Any] = (sample_prev - pred_original_sample) / sigma_prev A_ : int = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowercase , derivative=lowercase , state=lowercase ) def _a (self , lowercase , lowercase , lowercase , lowercase ): raise NotImplementedError()
206
'''simple docstring''' class _lowerCAmelCase ( __UpperCAmelCase ): pass class _lowerCAmelCase ( __UpperCAmelCase ): pass class _lowerCAmelCase : def __init__(self ): A_ : List[Any] = [ [], [], [], ] def _a (self , lowercase , lowercase ): try: if len(self.queues[priority] ) >= 100: raise OverflowError("""Maximum queue size is 100""" ) self.queues[priority].append(lowercase ) except IndexError: raise ValueError("""Valid priorities are 0, 1, and 2""" ) def _a (self ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("""All queues are empty""" ) def __str__(self ): return "\n".join(F'Priority {i}: {q}' for i, q in enumerate(self.queues ) ) class _lowerCAmelCase : def __init__(self ): A_ : List[str] = [] def _a (self , lowercase ): if len(self.queue ) == 100: raise OverFlowError("""Maximum queue size is 100""" ) self.queue.append(lowercase ) def _a (self ): if not self.queue: raise UnderFlowError("""The queue is empty""" ) else: A_ : List[Any] = min(self.queue ) self.queue.remove(lowercase ) return data def __str__(self ): return str(self.queue ) def a ( ): '''simple docstring''' A_ : Optional[int] = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 1_00 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 1_28 ) print(lowerCamelCase__ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(lowerCamelCase__ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def a ( ): '''simple docstring''' A_ : int = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(1_00 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(1_28 ) print(lowerCamelCase__ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(lowerCamelCase__ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
206
1
import math import random def lowercase( UpperCamelCase_ , UpperCamelCase_ = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _SCREAMING_SNAKE_CASE = 0.02 def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> float: '''simple docstring''' UpperCamelCase = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(UpperCamelCase_ ): # Forward propagation UpperCamelCase = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCamelCase = (expected / 100) - layer_a # Error delta UpperCamelCase = layer_1_error * sigmoid_function(UpperCamelCase_ , UpperCamelCase_ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = int(input("""Expected value: """)) _SCREAMING_SNAKE_CASE = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
165
def lowercase( UpperCamelCase_ ) -> list[list]: '''simple docstring''' UpperCamelCase = current_set.copy() for row_index, row in enumerate(UpperCamelCase_ ): UpperCamelCase = row[0] for column_index, column in enumerate(UpperCamelCase_ ): if magnitude == 0: UpperCamelCase = column continue UpperCamelCase = column / magnitude # Subtract to cancel term UpperCamelCase = current_set[0] UpperCamelCase = [first_row] UpperCamelCase = current_set[1::] for row in current_set: UpperCamelCase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(UpperCamelCase_ ) continue for column_index in range(len(UpperCamelCase_ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(UpperCamelCase_ ) # Create next recursion iteration set if len(final_set[0] ) != 3: UpperCamelCase = final_set[0] UpperCamelCase = [] UpperCamelCase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) UpperCamelCase = simplify(UpperCamelCase_ ) for i in range(len(UpperCamelCase_ ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , UpperCamelCase_ ) UpperCamelCase = resultant return final_set def lowercase( UpperCamelCase_ ) -> list: '''simple docstring''' if len(UpperCamelCase_ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) UpperCamelCase = len(UpperCamelCase_ ) + 1 if any(len(UpperCamelCase_ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(UpperCamelCase_ , (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(UpperCamelCase_ ) == 1: return [equations[0][-1] / equations[0][0]] UpperCamelCase = equations.copy() if any(0 in row for row in data_set ): UpperCamelCase = data_set.copy() UpperCamelCase = [] for row_index, row in enumerate(UpperCamelCase_ ): if 0 not in row: UpperCamelCase = data_set.pop(UpperCamelCase_ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0 , UpperCamelCase_ ) UpperCamelCase = data_set.copy() UpperCamelCase = simplify(UpperCamelCase_ ) UpperCamelCase = simplified[::-1] UpperCamelCase = [] for row in simplified: UpperCamelCase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue UpperCamelCase = row.copy()[: len(UpperCamelCase_ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(UpperCamelCase_ ) == 0: solutions.append(0 ) continue UpperCamelCase = temp_row[1::] UpperCamelCase = temp_row[::-1] for column_index, column in enumerate(UpperCamelCase_ ): current_solution -= column * solutions[column_index] solutions.append(UpperCamelCase_ ) UpperCamelCase = [] for item in solutions: final.append(float(round(UpperCamelCase_ , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
165
1
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker __lowercase = '''CompVis/stable-diffusion-v1-1''' __lowercase = '''CompVis/stable-diffusion-v1-2''' __lowercase = '''CompVis/stable-diffusion-v1-3''' __lowercase = '''CompVis/stable-diffusion-v1-4''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = True , ) -> Union[str, Any]: super()._init_() __UpperCamelCase :Dict = StableDiffusionPipeline.from_pretrained(__lowercase) __UpperCamelCase :Union[str, Any] = StableDiffusionPipeline.from_pretrained(__lowercase) __UpperCamelCase :Optional[int] = StableDiffusionPipeline.from_pretrained(__lowercase) __UpperCamelCase :Tuple = StableDiffusionPipeline( vae=__lowercase , text_encoder=__lowercase , tokenizer=__lowercase , unet=__lowercase , scheduler=__lowercase , safety_checker=__lowercase , feature_extractor=__lowercase , requires_safety_checker=__lowercase , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea) @property def UpperCamelCase__ ( self) -> Dict[str, Any]: return {k: getattr(self , __lowercase) for k in self.config.keys() if not k.startswith('''_''')} def UpperCamelCase__ ( self , __lowercase = "auto") -> str: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __UpperCamelCase :Optional[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__lowercase) def UpperCamelCase__ ( self) -> str: self.enable_attention_slicing(__lowercase) @torch.no_grad() def UpperCamelCase__ ( self , __lowercase , __lowercase = 512 , __lowercase = 512 , __lowercase = 50 , __lowercase = 7.5 , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = None , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = None , __lowercase = 1 , **__lowercase , ) -> Union[str, Any]: return self.pipea( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) @torch.no_grad() def UpperCamelCase__ ( self , __lowercase , __lowercase = 512 , __lowercase = 512 , __lowercase = 50 , __lowercase = 7.5 , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = None , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = None , __lowercase = 1 , **__lowercase , ) -> int: return self.pipea( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) @torch.no_grad() def UpperCamelCase__ ( self , __lowercase , __lowercase = 512 , __lowercase = 512 , __lowercase = 50 , __lowercase = 7.5 , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = None , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = None , __lowercase = 1 , **__lowercase , ) -> str: return self.pipea( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) @torch.no_grad() def UpperCamelCase__ ( self , __lowercase , __lowercase = 512 , __lowercase = 512 , __lowercase = 50 , __lowercase = 7.5 , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = None , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = None , __lowercase = 1 , **__lowercase , ) -> str: return self.pipea( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) @torch.no_grad() def UpperCamelCase__ ( self , __lowercase , __lowercase = 512 , __lowercase = 512 , __lowercase = 50 , __lowercase = 7.5 , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = None , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = None , __lowercase = 1 , **__lowercase , ) -> Optional[int]: __UpperCamelCase :Optional[Any] = '''cuda''' if torch.cuda.is_available() else '''cpu''' self.to(__lowercase) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` must be divisible by 8 but are {height} and {width}.""") # Get first result from Stable Diffusion Checkpoint v1.1 __UpperCamelCase :Tuple = self.textaimg_sda_a( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.2 __UpperCamelCase :Any = self.textaimg_sda_a( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.3 __UpperCamelCase :Dict = self.textaimg_sda_a( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.4 __UpperCamelCase :Optional[Any] = self.textaimg_sda_a( prompt=__lowercase , height=__lowercase , width=__lowercase , num_inference_steps=__lowercase , guidance_scale=__lowercase , negative_prompt=__lowercase , num_images_per_prompt=__lowercase , eta=__lowercase , generator=__lowercase , latents=__lowercase , output_type=__lowercase , return_dict=__lowercase , callback=__lowercase , callback_steps=__lowercase , **__lowercase , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]])
43
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ "transfo-xl-wt103": "https://huggingface.co/transfo-xl-wt103/resolve/main/config.json", } class a_ ( lowerCamelCase_ ): """simple docstring""" __UpperCAmelCase = 'transfo-xl' __UpperCAmelCase = ['mems'] __UpperCAmelCase = { 'n_token': 'vocab_size', 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] ,snake_case : List[Any]=267735 ,snake_case : Optional[int]=[20000, 40000, 200000] ,snake_case : int=1024 ,snake_case : Optional[Any]=1024 ,snake_case : Tuple=16 ,snake_case : int=64 ,snake_case : Union[str, Any]=4096 ,snake_case : List[str]=4 ,snake_case : int=False ,snake_case : int=18 ,snake_case : Tuple=1600 ,snake_case : List[str]=1000 ,snake_case : Optional[Any]=True ,snake_case : List[str]=True ,snake_case : Optional[Any]=0 ,snake_case : Optional[Any]=-1 ,snake_case : List[Any]=True ,snake_case : Optional[Any]=0.1 ,snake_case : Union[str, Any]=0.0 ,snake_case : int=True ,snake_case : Any="normal" ,snake_case : int=0.01 ,snake_case : int=0.01 ,snake_case : str=0.02 ,snake_case : Any=1e-5 ,snake_case : Optional[int]=0 ,**snake_case : List[Any] ,): SCREAMING_SNAKE_CASE =vocab_size SCREAMING_SNAKE_CASE =[] self.cutoffs.extend(snake_case ) if proj_share_all_but_first: SCREAMING_SNAKE_CASE =[False] + [True] * len(self.cutoffs ) else: SCREAMING_SNAKE_CASE =[False] + [False] * len(self.cutoffs ) SCREAMING_SNAKE_CASE =d_model SCREAMING_SNAKE_CASE =d_embed SCREAMING_SNAKE_CASE =d_head SCREAMING_SNAKE_CASE =d_inner SCREAMING_SNAKE_CASE =div_val SCREAMING_SNAKE_CASE =pre_lnorm SCREAMING_SNAKE_CASE =n_layer SCREAMING_SNAKE_CASE =n_head SCREAMING_SNAKE_CASE =mem_len SCREAMING_SNAKE_CASE =same_length SCREAMING_SNAKE_CASE =attn_type SCREAMING_SNAKE_CASE =clamp_len SCREAMING_SNAKE_CASE =sample_softmax SCREAMING_SNAKE_CASE =adaptive SCREAMING_SNAKE_CASE =dropout SCREAMING_SNAKE_CASE =dropatt SCREAMING_SNAKE_CASE =untie_r SCREAMING_SNAKE_CASE =init SCREAMING_SNAKE_CASE =init_range SCREAMING_SNAKE_CASE =proj_init_std SCREAMING_SNAKE_CASE =init_std SCREAMING_SNAKE_CASE =layer_norm_epsilon super().__init__(eos_token_id=snake_case ,**snake_case ) @property def _lowerCAmelCase ( self : str ): # Message copied from Transformer-XL documentation logger.info(f'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Dict ): # Message copied from Transformer-XL documentation raise NotImplementedError( f'The model {self.model_type} is one of the few models that has no sequence length limit.' )
334
0
'''simple docstring''' def lowercase__( __UpperCamelCase: int = 1_00_00_00 ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = set(range(3 ,__a ,2 ) ) primes.add(2 ) for p in range(3 ,__a ,2 ): if p not in primes: continue primes.difference_update(set(range(p * p ,__a ,__a ) ) ) SCREAMING_SNAKE_CASE : Any = [float(__a ) for n in range(limit + 1 )] for p in primes: for n in range(__a ,limit + 1 ,__a ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"""{solution() = }""")
365
'''simple docstring''' import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _a : '''simple docstring''' def __init__( self, A, A=2, A=3, A=4, A=2, A=7, A=True, A=True, A=True, A=True, A=99, A=36, A=3, A=4, A=37, A="gelu", A=0.1, A=0.1, A=512, A=16, A=2, A=0.02, A=6, A=6, A=3, A=4, A=None, A=1_000, ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = parent SCREAMING_SNAKE_CASE : List[Any] = batch_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : Optional[int] = image_size SCREAMING_SNAKE_CASE : List[str] = patch_size SCREAMING_SNAKE_CASE : List[Any] = text_seq_length SCREAMING_SNAKE_CASE : Tuple = is_training SCREAMING_SNAKE_CASE : Tuple = use_input_mask SCREAMING_SNAKE_CASE : Optional[int] = use_token_type_ids SCREAMING_SNAKE_CASE : Optional[Any] = use_labels SCREAMING_SNAKE_CASE : Optional[int] = vocab_size SCREAMING_SNAKE_CASE : Any = hidden_size SCREAMING_SNAKE_CASE : List[str] = num_hidden_layers SCREAMING_SNAKE_CASE : Union[str, Any] = num_attention_heads SCREAMING_SNAKE_CASE : Any = intermediate_size SCREAMING_SNAKE_CASE : List[Any] = hidden_act SCREAMING_SNAKE_CASE : Any = hidden_dropout_prob SCREAMING_SNAKE_CASE : Union[str, Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : str = max_position_embeddings SCREAMING_SNAKE_CASE : List[str] = type_vocab_size SCREAMING_SNAKE_CASE : Tuple = type_sequence_label_size SCREAMING_SNAKE_CASE : List[Any] = initializer_range SCREAMING_SNAKE_CASE : str = coordinate_size SCREAMING_SNAKE_CASE : Tuple = shape_size SCREAMING_SNAKE_CASE : Optional[Any] = num_labels SCREAMING_SNAKE_CASE : int = num_choices SCREAMING_SNAKE_CASE : str = scope SCREAMING_SNAKE_CASE : Optional[int] = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) SCREAMING_SNAKE_CASE : Union[str, Any] = text_seq_length SCREAMING_SNAKE_CASE : List[str] = (image_size // patch_size) ** 2 + 1 SCREAMING_SNAKE_CASE : int = self.text_seq_length + self.image_seq_length def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = ids_tensor([self.batch_size, self.text_seq_length], self.vocab_size ) SCREAMING_SNAKE_CASE : int = ids_tensor([self.batch_size, self.text_seq_length, 4], self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: SCREAMING_SNAKE_CASE : str = bbox[i, j, 3] SCREAMING_SNAKE_CASE : Optional[Any] = bbox[i, j, 1] SCREAMING_SNAKE_CASE : Tuple = t if bbox[i, j, 2] < bbox[i, j, 0]: SCREAMING_SNAKE_CASE : Optional[Any] = bbox[i, j, 2] SCREAMING_SNAKE_CASE : Any = bbox[i, j, 0] SCREAMING_SNAKE_CASE : Tuple = t SCREAMING_SNAKE_CASE : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : Any = None if self.use_input_mask: SCREAMING_SNAKE_CASE : List[Any] = random_attention_mask([self.batch_size, self.text_seq_length] ) SCREAMING_SNAKE_CASE : List[Any] = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE : List[Any] = ids_tensor([self.batch_size, self.text_seq_length], self.type_vocab_size ) SCREAMING_SNAKE_CASE : Optional[Any] = None SCREAMING_SNAKE_CASE : Tuple = None if self.use_labels: SCREAMING_SNAKE_CASE : int = ids_tensor([self.batch_size], self.type_sequence_label_size ) SCREAMING_SNAKE_CASE : Dict = ids_tensor([self.batch_size, self.text_seq_length], self.num_labels ) SCREAMING_SNAKE_CASE : int = LayoutLMvaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, coordinate_size=self.coordinate_size, shape_size=self.shape_size, input_size=self.image_size, patch_size=self.patch_size, ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def UpperCamelCase_ ( self, A, A, A, A, A, A, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = LayoutLMvaModel(config=A ) model.to(A ) model.eval() # text + image SCREAMING_SNAKE_CASE : Optional[int] = model(A, pixel_values=A ) SCREAMING_SNAKE_CASE : Union[str, Any] = model( A, bbox=A, pixel_values=A, attention_mask=A, token_type_ids=A ) SCREAMING_SNAKE_CASE : List[str] = model(A, bbox=A, pixel_values=A, token_type_ids=A ) SCREAMING_SNAKE_CASE : Optional[int] = model(A, bbox=A, pixel_values=A ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) # text only SCREAMING_SNAKE_CASE : List[Any] = model(A ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only SCREAMING_SNAKE_CASE : List[str] = model(pixel_values=A ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.image_seq_length, self.hidden_size) ) def UpperCamelCase_ ( self, A, A, A, A, A, A, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = self.num_labels SCREAMING_SNAKE_CASE : List[str] = LayoutLMvaForSequenceClassification(A ) model.to(A ) model.eval() SCREAMING_SNAKE_CASE : Dict = model( A, bbox=A, pixel_values=A, attention_mask=A, token_type_ids=A, labels=A, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self, A, A, A, A, A, A, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = self.num_labels SCREAMING_SNAKE_CASE : str = LayoutLMvaForTokenClassification(config=A ) model.to(A ) model.eval() SCREAMING_SNAKE_CASE : str = model( A, bbox=A, pixel_values=A, attention_mask=A, token_type_ids=A, labels=A, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.text_seq_length, self.num_labels) ) def UpperCamelCase_ ( self, A, A, A, A, A, A, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = LayoutLMvaForQuestionAnswering(config=A ) model.to(A ) model.eval() SCREAMING_SNAKE_CASE : Optional[int] = model( A, bbox=A, pixel_values=A, attention_mask=A, token_type_ids=A, start_positions=A, end_positions=A, ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ) : Any = config_and_inputs SCREAMING_SNAKE_CASE : Dict = { 'input_ids': input_ids, 'bbox': bbox, 'pixel_values': pixel_values, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' A : Optional[int] = False A : List[str] = False A : Union[str, Any] = False A : Optional[Any] = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) A : List[Any] = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def UpperCamelCase_ ( self, A, A, A, A, A ): '''simple docstring''' return True def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = LayoutLMvaModelTester(self ) SCREAMING_SNAKE_CASE : Optional[Any] = ConfigTester(self, config_class=A, hidden_size=37 ) def UpperCamelCase_ ( self, A, A, A=False ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = copy.deepcopy(A ) if model_class in get_values(A ): SCREAMING_SNAKE_CASE : Optional[int] = { k: v.unsqueeze(1 ).expand(-1, self.model_tester.num_choices, -1 ).contiguous() if isinstance(A, torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(A ): SCREAMING_SNAKE_CASE : Optional[Any] = torch.ones(self.model_tester.batch_size, dtype=torch.long, device=A ) elif model_class in get_values(A ): SCREAMING_SNAKE_CASE : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=A ) SCREAMING_SNAKE_CASE : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=A ) elif model_class in [ *get_values(A ), ]: SCREAMING_SNAKE_CASE : str = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=A ) elif model_class in [ *get_values(A ), ]: SCREAMING_SNAKE_CASE : Union[str, Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length), dtype=torch.long, device=A, ) return inputs_dict def UpperCamelCase_ ( self ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE : List[str] = type self.model_tester.create_and_check_model(*A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE : Optional[Any] = LayoutLMvaModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase__( ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class _a ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ): '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=A ) if is_vision_available() else None @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = LayoutLMvaModel.from_pretrained('microsoft/layoutlmv3-base' ).to(A ) SCREAMING_SNAKE_CASE : Any = self.default_image_processor SCREAMING_SNAKE_CASE : Any = prepare_img() SCREAMING_SNAKE_CASE : Dict = image_processor(images=A, return_tensors='pt' ).pixel_values.to(A ) SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor([[1, 2]] ) SCREAMING_SNAKE_CASE : Tuple = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass SCREAMING_SNAKE_CASE : Union[str, Any] = model( input_ids=input_ids.to(A ), bbox=bbox.to(A ), pixel_values=pixel_values.to(A ), ) # verify the logits SCREAMING_SNAKE_CASE : str = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape, A ) SCREAMING_SNAKE_CASE : Tuple = torch.tensor( [[-0.05_29, 0.36_18, 0.16_32], [-0.15_87, -0.16_67, -0.04_00], [-0.15_57, -0.16_71, -0.05_05]] ).to(A ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3], A, atol=1E-4 ) )
246
0
def A (__A : list[int] ) -> float: """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) UpperCAmelCase_ = sum(__A ) / len(__A ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(__A ) if __name__ == "__main__": import doctest doctest.testmod()
51
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=_a ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) snake_case__ : ClassVar[Features] = Features({"""audio""": Audio()} ) snake_case__ : ClassVar[Features] = Features({"""transcription""": Value("""string""" )} ) snake_case__ : str = "audio" snake_case__ : str = "transcription" def _A ( self : List[str] , __lowerCamelCase : Dict ): if self.audio_column not in features: raise ValueError(F"""Column {self.audio_column} is not present in features.""" ) if not isinstance(features[self.audio_column] , __lowerCamelCase ): raise ValueError(F"""Column {self.audio_column} is not an Audio type.""" ) UpperCamelCase :int = copy.deepcopy(self ) UpperCamelCase :Any = self.input_schema.copy() UpperCamelCase :List[str] = features[self.audio_column] UpperCamelCase :List[Any] = input_schema return task_template @property def _A ( self : Optional[int] ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
38
0
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Any = inspect.getfile(accelerate.test_utils) __UpperCamelCase :Any = os.path.sep.join( mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py''']) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 __UpperCamelCase :List[str] = test_metrics @require_cpu def UpperCamelCase__ ( self) -> List[Any]: debug_launcher(self.test_metrics.main , num_processes=1) @require_cpu def UpperCamelCase__ ( self) -> Optional[int]: debug_launcher(self.test_metrics.main) @require_single_gpu def UpperCamelCase__ ( self) -> Union[str, Any]: self.test_metrics.main() @require_multi_gpu def UpperCamelCase__ ( self) -> List[Any]: print(f"""Found {torch.cuda.device_count()} devices.""") __UpperCamelCase :Union[str, Any] = ['''torchrun''', f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1): execute_subprocess_async(__lowercase , env=os.environ.copy())
105
import json import os import tempfile from unittest.mock import patch import torch from torch.utils.data import DataLoader, TensorDataset from accelerate import DistributedType, infer_auto_device_map, init_empty_weights from accelerate.accelerator import Accelerator from accelerate.state import GradientState, PartialState from accelerate.test_utils import require_bnb, require_multi_gpu, slow from accelerate.test_utils.testing import AccelerateTestCase, require_cuda from accelerate.utils import patch_environment def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :List[str] = torch.nn.Linear(2 , 4 ) __UpperCamelCase :Any = torch.optim.AdamW(model.parameters() , lr=1.0 ) __UpperCamelCase :List[Any] = torch.optim.lr_scheduler.OneCycleLR(SCREAMING_SNAKE_CASE , max_lr=0.01 , steps_per_epoch=2 , epochs=1 ) __UpperCamelCase :List[Any] = DataLoader(TensorDataset(torch.tensor([1, 2, 3] ) ) ) __UpperCamelCase :Dict = DataLoader(TensorDataset(torch.tensor([4, 5, 6] ) ) ) return model, optimizer, scheduler, train_dl, valid_dl def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return (model.weight.abs().sum() + model.bias.abs().sum()).item() def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = torch.nn.Linear(*tuple(model.weight.T.shape ) ).state_dict() model.load_state_dict(SCREAMING_SNAKE_CASE ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' @require_cuda def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Dict = Accelerator() assert PartialState._shared_state["_cpu"] is False assert PartialState._shared_state["device"].type == "cuda" with self.assertRaises(__lowercase): __UpperCamelCase :Any = Accelerator(cpu=__lowercase) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :List[Any] = Accelerator() __UpperCamelCase :List[Any] = GradientState() assert state.num_steps == 1 __UpperCamelCase :Any = 4 assert state.num_steps == 4 assert state.sync_gradients is True __UpperCamelCase :int = False assert state.sync_gradients is False GradientState._reset_state() def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Tuple = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[Any] = create_components() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) :int = accelerator.prepare(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase) self.assertTrue(prepared_model in accelerator._models) self.assertTrue(prepared_optimizer in accelerator._optimizers) self.assertTrue(prepared_scheduler in accelerator._schedulers) self.assertTrue(prepared_train_dl in accelerator._dataloaders) self.assertTrue(prepared_valid_dl in accelerator._dataloaders) def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :str = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[Any] = create_components() accelerator.prepare(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase) accelerator.free_memory() self.assertTrue(len(accelerator._models) == 0) self.assertTrue(len(accelerator._optimizers) == 0) self.assertTrue(len(accelerator._schedulers) == 0) self.assertTrue(len(accelerator._dataloaders) == 0) def UpperCamelCase__ ( self) -> Union[str, Any]: PartialState._reset_state() # Mock torch.cuda.set_device to avoid an exception as the device doesn't exist def noop(*__lowercase , **__lowercase): pass with patch('''torch.cuda.set_device''' , __lowercase), patch_environment(ACCELERATE_TORCH_DEVICE='''cuda:64'''): __UpperCamelCase :Optional[Any] = Accelerator() self.assertEqual(str(accelerator.state.device) , '''cuda:64''') def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :List[Any] = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[Any] = create_components() accelerator.prepare(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase) __UpperCamelCase :Tuple = get_signature(__lowercase) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(__lowercase) # make sure random weights don't match load_random_weights(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) > 1E-3) # make sure loaded weights match accelerator.load_state(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) < 1E-3) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :List[Any] = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = create_components() accelerator.prepare(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase) __UpperCamelCase :Any = get_signature(__lowercase) # saving hook def save_config(__lowercase , __lowercase , __lowercase): __UpperCamelCase :Union[str, Any] = {'''class_name''': models[0].__class__.__name__} with open(os.path.join(__lowercase , '''data.json''') , '''w''') as f: json.dump(__lowercase , __lowercase) # loading hook def load_config(__lowercase , __lowercase): with open(os.path.join(__lowercase , '''data.json''') , '''r''') as f: __UpperCamelCase :Dict = json.load(__lowercase) __UpperCamelCase :Dict = config['''class_name'''] __UpperCamelCase :Union[str, Any] = accelerator.register_save_state_pre_hook(__lowercase) __UpperCamelCase :Any = accelerator.register_load_state_pre_hook(__lowercase) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(__lowercase) # make sure random weights don't match with hooks load_random_weights(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) > 1E-3) # random class name to verify correct one is loaded __UpperCamelCase :int = '''random''' # make sure loaded weights match with hooks accelerator.load_state(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) < 1E-3) # mode.class_name is loaded from config self.assertTrue(model.class_name == model.__class__.__name__) # remove hooks save_hook.remove() load_hook.remove() with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(__lowercase) # make sure random weights don't match with hooks removed load_random_weights(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) > 1E-3) # random class name to verify correct one is loaded __UpperCamelCase :Dict = '''random''' # make sure loaded weights match with hooks removed accelerator.load_state(__lowercase) self.assertTrue(abs(model_signature - get_signature(__lowercase)) < 1E-3) # mode.class_name is NOT loaded from config self.assertTrue(model.class_name != model.__class__.__name__) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Optional[Any] = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Union[str, Any] = create_components() __UpperCamelCase :Optional[Any] = None # This should work __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.prepare( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) self.assertTrue(dummy_obj is None) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[str] = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = create_components() __UpperCamelCase :Dict = [1, 2, 3] # This should work __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Tuple = accelerator.prepare( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Dummy object should have `_is_accelerate_prepared` set to `True`''' , ) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Model is missing `_is_accelerator_prepared` or is set to `False`''' , ) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Optimizer is missing `_is_accelerator_prepared` or is set to `False`''' , ) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Scheduler is missing `_is_accelerator_prepared` or is set to `False`''' , ) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Train Dataloader is missing `_is_accelerator_prepared` or is set to `False`''' , ) self.assertEqual( getattr(__lowercase , '''_is_accelerate_prepared''' , __lowercase) , __lowercase , '''Valid Dataloader is missing `_is_accelerator_prepared` or is set to `False`''' , ) @slow @require_bnb def UpperCamelCase__ ( self) -> int: from transformers import AutoModelForCausalLM __UpperCamelCase :Optional[int] = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , load_in_abit=__lowercase , device_map={'''''': 0} , ) __UpperCamelCase :Optional[Any] = Accelerator() # This should work __UpperCamelCase :int = accelerator.prepare(__lowercase) @slow @require_bnb def UpperCamelCase__ ( self) -> List[str]: from transformers import AutoModelForCausalLM __UpperCamelCase :str = Accelerator() with init_empty_weights(): __UpperCamelCase :List[Any] = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , ) model.tie_weights() __UpperCamelCase :List[str] = infer_auto_device_map(__lowercase) __UpperCamelCase :str = '''cpu''' __UpperCamelCase :List[Any] = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , device_map=__lowercase , load_in_abit=__lowercase , llm_inta_enable_fpaa_cpu_offload=__lowercase) # This should not work and get value error with self.assertRaises(__lowercase): __UpperCamelCase :Union[str, Any] = accelerator.prepare(__lowercase) @slow @require_bnb @require_multi_gpu def UpperCamelCase__ ( self) -> Union[str, Any]: from transformers import AutoModelForCausalLM __UpperCamelCase :int = {'''distributed_type''': DistributedType.MULTI_GPU} with init_empty_weights(): __UpperCamelCase :Tuple = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , ) model.tie_weights() __UpperCamelCase :int = infer_auto_device_map(__lowercase) __UpperCamelCase :List[Any] = 1 __UpperCamelCase :int = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , load_in_abit=__lowercase , device_map=__lowercase , ) __UpperCamelCase :Dict = Accelerator() # This should not work and get value error with self.assertRaises(__lowercase): __UpperCamelCase :Any = accelerator.prepare(__lowercase) PartialState._reset_state() @slow @require_bnb @require_multi_gpu def UpperCamelCase__ ( self) -> Dict: from transformers import AutoModelForCausalLM with init_empty_weights(): __UpperCamelCase :Optional[int] = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , ) __UpperCamelCase :List[str] = infer_auto_device_map(__lowercase) __UpperCamelCase :Optional[int] = 1 __UpperCamelCase :Optional[int] = AutoModelForCausalLM.from_pretrained( '''EleutherAI/gpt-neo-125m''' , load_in_abit=__lowercase , device_map=__lowercase , ) __UpperCamelCase :int = Accelerator() # This should work __UpperCamelCase :int = accelerator.prepare(__lowercase) @require_cuda def UpperCamelCase__ ( self) -> int: __UpperCamelCase :Tuple = torch.nn.Linear(10 , 10) __UpperCamelCase :Optional[Any] = torch.optim.SGD(model.parameters() , lr=0.01) __UpperCamelCase :Any = Accelerator(cpu=__lowercase) __UpperCamelCase :Tuple = accelerator.prepare(__lowercase)
105
1
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy A_ : int = logging.getLogger(__name__) A_ : Tuple = "pytorch_model.bin" @dataclasses.dataclass class _a : '''simple docstring''' UpperCAmelCase__: Union[str, Any] = dataclasses.field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models.'''} ) UpperCAmelCase__: Optional[Any] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co.'''} , ) @dataclasses.dataclass class _a : '''simple docstring''' UpperCAmelCase__: Dict = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the training data.'''} ) UpperCAmelCase__: List[str] = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the data to predict on.'''} ) UpperCAmelCase__: List[Any] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''A csv or a json file containing the validation data.'''} ) UpperCAmelCase__: Optional[int] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''The name of the task to train on.'''} , ) UpperCAmelCase__: Optional[Any] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''The list of labels for the task.'''} ) @dataclasses.dataclass class _a : '''simple docstring''' UpperCAmelCase__: List[str] = dataclasses.field( metadata={'''help''': '''The output directory where the model predictions and checkpoints will be written.'''} ) UpperCAmelCase__: Optional[int] = dataclasses.field( default='''accuracy''' , metadata={'''help''': '''The evaluation metric used for the task.'''} ) UpperCAmelCase__: int = dataclasses.field( default='''no''' , metadata={ '''help''': '''The evaluation strategy to adopt during training. Possible values are: ["no", "step", "epoch]''' } , ) UpperCAmelCase__: Optional[Any] = dataclasses.field( default=10 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) UpperCAmelCase__: List[Any] = dataclasses.field( default=0.0 , metadata={ '''help''': '''How much the specified evaluation metric must improve to satisfy early stopping conditions.''' } , ) UpperCAmelCase__: Union[str, Any] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the confidence score.'''} , ) UpperCAmelCase__: Optional[Any] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the validation performance.'''} , ) UpperCAmelCase__: Dict = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''Whether to fine-tune on labeled data after pseudo training.'''} , ) UpperCAmelCase__: Optional[int] = dataclasses.field( default=0.0 , metadata={'''help''': '''Confidence threshold for pseudo-labeled data filtering.'''} , ) UpperCAmelCase__: int = dataclasses.field( default=1_00 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) UpperCAmelCase__: List[str] = dataclasses.field( default=UpperCAmelCase__ , metadata={'''help''': '''Random seed for initialization.'''} , ) def UpperCamelCase (lowercase_: Optional[int] , lowercase_: Optional[int] , lowercase_: Any , lowercase_: Union[str, Any] , lowercase_: str , lowercase_: int ) -> int: A__ : List[str] = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: A__ : Any = dataset.filter(lambda lowercase_ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 A__ : Optional[Any] = int(eval_result * len(_A ) ) print(_A ) A__ : Optional[Any] = dataset.sort("""probability""" , reverse=_A ) A__ : List[Any] = dataset.select(range(_A ) ) A__ : Dict = dataset.remove_columns(["""label""", """probability"""] ) A__ : int = dataset.rename_column("""prediction""" , """label""" ) A__ : int = dataset.map(lambda lowercase_ : {"label": idalabel[example["label"]]} ) A__ : Union[str, Any] = dataset.shuffle(seed=args.seed ) A__ : Tuple = os.path.join(_A , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(_A , index=_A ) else: dataset.to_json(_A ) def UpperCamelCase (lowercase_: List[str] , lowercase_: int , lowercase_: Optional[int] , lowercase_: Optional[int] , **lowercase_: Dict ) -> List[str]: A__ : List[Any] = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() A__ : Optional[Any] = STModelArguments(model_name_or_path=_A ) A__ : Union[str, Any] = STDataArguments(train_file=_A , infer_file=_A ) A__ : int = STTrainingArguments(output_dir=_A ) A__ : Union[str, Any] = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(_A ).items(): setattr(_A , _A , _A ) for key, value in kwargs.items(): if hasattr(_A , _A ): setattr(_A , _A , _A ) # Sanity checks A__ : Any = {} A__ : Tuple = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None A__ : Tuple = args.train_file A__ : Tuple = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None A__ : List[str] = args.eval_file for key in data_files: A__ : Any = data_files[key].split(""".""" )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: A__ : Dict = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("""Creating the initial data directory for self-training...""" ) A__ : Optional[Any] = f"""{args.output_dir}/self-train_iter-{{}}""".format A__ : Any = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=_A ) os.makedirs(_A , exist_ok=_A ) accelerator.wait_for_everyone() A__ : Optional[Any] = None A__ : str = None A__ : Union[str, Any] = 0 A__ : int = False # Show the progress bar A__ : List[Any] = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): A__ : Optional[Any] = data_dir_format(_A ) assert os.path.exists(_A ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 A__ : Union[str, Any] = os.path.join(_A , """stage-1""" ) A__ : List[Any] = { """accelerator""": accelerator, """model_name_or_path""": args.model_name_or_path, """cache_dir""": args.cache_dir, """do_train""": True, """train_file""": data_files["""train"""] if iteration == 0 else data_files["""train_pseudo"""], """do_eval""": True if args.eval_file is not None else False, """eval_file""": data_files["""eval"""], """do_predict""": True, """infer_file""": data_files["""infer"""], """task_name""": args.task_name, """label_list""": args.label_list, """output_dir""": current_output_dir, """eval_metric""": args.eval_metric, """evaluation_strategy""": args.evaluation_strategy, """early_stopping_patience""": args.early_stopping_patience, """early_stopping_threshold""": args.early_stopping_threshold, """seed""": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(_A , _A ): arguments_dict.update({key: value} ) A__ : List[Any] = os.path.join(_A , """best-checkpoint""" , _A ) if os.path.exists(_A ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.""" , _A , _A , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 1 *****""" , _A ) finetune(**_A ) accelerator.wait_for_everyone() assert os.path.exists(_A ) logger.info("""Self-training job completed: iteration: %d, stage: 1.""" , _A ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data A__ : Tuple = os.path.join(_A , """best-checkpoint""" ) A__ : List[str] = os.path.join(_A , """stage-2""" ) # Update arguments_dict A__ : str = model_path A__ : Optional[int] = data_files["""train"""] A__ : List[Any] = current_output_dir A__ : str = os.path.join(_A , """best-checkpoint""" , _A ) if os.path.exists(_A ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.""" , _A , _A , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 2 *****""" , _A ) finetune(**_A ) accelerator.wait_for_everyone() assert os.path.exists(_A ) logger.info("""Self-training job completed: iteration: %d, stage: 2.""" , _A ) A__ : Any = iteration A__ : Optional[int] = data_dir_format(iteration + 1 ) A__ : List[Any] = AutoConfig.from_pretrained(os.path.join(_A , """best-checkpoint""" ) ) A__ : Optional[int] = config.idalabel A__ : List[str] = os.path.join(_A , """eval_results_best-checkpoint.json""" ) A__ : str = os.path.join(_A , """test_results_best-checkpoint.json""" ) assert os.path.exists(_A ) with open(_A , """r""" ) as f: A__ : List[str] = float(json.load(_A )[args.eval_metric] ) A__ : Any = os.path.join(_A , """infer_output_best-checkpoint.csv""" ) assert os.path.exists(_A ) # Loading the dataset from local csv or json files. A__ : Optional[Any] = load_dataset(args.data_file_extension , data_files={"""data""": data_files["""infer"""]} )["""data"""] A__ : Optional[Any] = load_dataset("""csv""" , data_files={"""data""": infer_output_file} )["""data"""] if accelerator.is_main_process: os.makedirs(_A , exist_ok=_A ) shutil.copy(_A , os.path.join(_A , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(_A ): shutil.copy(_A , os.path.join(_A , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(_A , _A , _A , _A , _A , _A ) accelerator.wait_for_everyone() A__ : Union[str, Any] = os.path.join(_A , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: A__ : Any = eval_result if best_iteration is None: A__ : int = new_iteration A__ : List[Any] = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: A__ : Optional[int] = new_iteration A__ : Dict = new_eval_result A__ : Any = 0 else: if new_eval_result == best_eval_result: A__ : List[Any] = new_iteration A__ : Optional[Any] = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: A__ : Dict = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("""Best iteration: %d""" , _A ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , _A ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_A , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(_A , """eval_results_best-iteration.json""" ) , ) else: # Assume that the last iteration is the best logger.info("""Best iteration: %d""" , args.max_selftrain_iterations - 1 ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , _A ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_A , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(_A , """eval_results_best-iteration.json""" ) , )
192
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = None if token is not None: snake_case_ = {"Accept": "application/vnd.github+json", "Authorization": f"Bearer {token}"} snake_case_ = f"https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100" snake_case_ = requests.get(_A , headers=_A ).json() snake_case_ = {} try: job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) snake_case_ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(_A ): snake_case_ = requests.get(url + f"&page={i + 2}" , headers=_A ).json() job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) return job_links except Exception: print(f"Unknown error, could not fetch links:\n{traceback.format_exc()}" ) return {} def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = None if token is not None: snake_case_ = {"Accept": "application/vnd.github+json", "Authorization": f"Bearer {token}"} snake_case_ = f"https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100" snake_case_ = requests.get(_A , headers=_A ).json() snake_case_ = {} try: artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) snake_case_ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(_A ): snake_case_ = requests.get(url + f"&page={i + 2}" , headers=_A ).json() artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) return artifacts except Exception: print(f"Unknown error, could not fetch links:\n{traceback.format_exc()}" ) return {} def lowerCamelCase__ ( _A , _A , _A , _A ): '''simple docstring''' snake_case_ = None if token is not None: snake_case_ = {"Accept": "application/vnd.github+json", "Authorization": f"Bearer {token}"} snake_case_ = requests.get(_A , headers=_A , allow_redirects=_A ) snake_case_ = result.headers["Location"] snake_case_ = requests.get(_A , allow_redirects=_A ) snake_case_ = os.path.join(_A , f"{artifact_name}.zip" ) with open(_A , "wb" ) as fp: fp.write(response.content ) def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = [] snake_case_ = [] snake_case_ = None with zipfile.ZipFile(_A ) as z: for filename in z.namelist(): if not os.path.isdir(_A ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_A ) as f: for line in f: snake_case_ = line.decode("UTF-8" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs snake_case_ = line[: line.index(": " )] snake_case_ = line[line.index(": " ) + len(": " ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("FAILED " ): # `test` is the test method that failed snake_case_ = line[len("FAILED " ) :] failed_tests.append(_A ) elif filename == "job_name.txt": snake_case_ = line if len(_A ) != len(_A ): raise ValueError( f"`errors` and `failed_tests` should have the same number of elements. Got {len(_A )} for `errors` " f"and {len(_A )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some" " problem." ) snake_case_ = None if job_name and job_links: snake_case_ = job_links.get(_A , _A ) # A list with elements of the form (line of error, error, failed test) snake_case_ = [x + [y] + [job_link] for x, y in zip(_A , _A )] return result def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = [] snake_case_ = [os.path.join(_A , _A ) for p in os.listdir(_A ) if p.endswith(".zip" )] for p in paths: errors.extend(get_errors_from_single_artifact(_A , job_links=_A ) ) return errors def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = Counter() counter.update([x[1] for x in logs] ) snake_case_ = counter.most_common() snake_case_ = {} for error, count in counts: if error_filter is None or error not in error_filter: snake_case_ = {"count": count, "failed_tests": [(x[2], x[0]) for x in logs if x[1] == error]} snake_case_ = dict(sorted(r.items() , key=lambda _A : item[1]["count"] , reverse=_A ) ) return r def lowerCamelCase__ ( _A ): '''simple docstring''' snake_case_ = test.split("::" )[0] if test.startswith("tests/models/" ): snake_case_ = test.split("/" )[2] else: snake_case_ = None return test def lowerCamelCase__ ( _A , _A=None ): '''simple docstring''' snake_case_ = [(x[0], x[1], get_model(x[2] )) for x in logs] snake_case_ = [x for x in logs if x[2] is not None] snake_case_ = {x[2] for x in logs} snake_case_ = {} for test in tests: snake_case_ = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) snake_case_ = counter.most_common() snake_case_ = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} snake_case_ = sum(error_counts.values() ) if n_errors > 0: snake_case_ = {"count": n_errors, "errors": error_counts} snake_case_ = dict(sorted(r.items() , key=lambda _A : item[1]["count"] , reverse=_A ) ) return r def lowerCamelCase__ ( _A ): '''simple docstring''' snake_case_ = "| no. | error | status |" snake_case_ = "|-:|:-|:-|" snake_case_ = [header, sep] for error in reduced_by_error: snake_case_ = reduced_by_error[error]["count"] snake_case_ = f"| {count} | {error[:100]} | |" lines.append(_A ) return "\n".join(_A ) def lowerCamelCase__ ( _A ): '''simple docstring''' snake_case_ = "| model | no. of errors | major error | count |" snake_case_ = "|-:|-:|-:|-:|" snake_case_ = [header, sep] for model in reduced_by_model: snake_case_ = reduced_by_model[model]["count"] snake_case_ , snake_case_ = list(reduced_by_model[model]["errors"].items() )[0] snake_case_ = f"| {model} | {count} | {error[:60]} | {_count} |" lines.append(_A ) return "\n".join(_A ) if __name__ == "__main__": lowercase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") parser.add_argument( "--output_dir", type=str, required=True, help="Where to store the downloaded artifacts and other result files.", ) parser.add_argument("--token", default=None, type=str, help="A token that has actions:read permission.") lowercase__ : int = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ : Optional[Any] = get_job_links(args.workflow_run_id, token=args.token) lowercase__ : List[Any] = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ : Tuple = k.find(" / ") lowercase__ : Union[str, Any] = k[index + len(" / ") :] lowercase__ : Union[str, Any] = v with open(os.path.join(args.output_dir, "job_links.json"), "w", encoding="UTF-8") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ : Any = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, "artifacts.json"), "w", encoding="UTF-8") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ : List[Any] = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ : str = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ : Union[str, Any] = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, "errors.json"), "w", encoding="UTF-8") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ : int = reduce_by_error(errors) lowercase__ : List[str] = reduce_by_model(errors) lowercase__ : Tuple = make_github_table(reduced_by_error) lowercase__ : Optional[int] = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, "reduced_by_error.txt"), "w", encoding="UTF-8") as fp: fp.write(sa) with open(os.path.join(args.output_dir, "reduced_by_model.txt"), "w", encoding="UTF-8") as fp: fp.write(sa)
187
0
from string import ascii_uppercase UpperCAmelCase : List[str] = {char: i for i, char in enumerate(ascii_uppercase)} UpperCAmelCase : Optional[int] = dict(enumerate(ascii_uppercase)) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = len(lowerCAmelCase__ ) lowercase_ = 0 while True: if x == i: lowercase_ = 0 if len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ): break key += key[i] i += 1 return key def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = """""" lowercase_ = 0 for letter in message: if letter == " ": cipher_text += " " else: lowercase_ = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = """""" lowercase_ = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: lowercase_ = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def _SCREAMING_SNAKE_CASE () -> None: '''simple docstring''' lowercase_ = """THE GERMAN ATTACK""" lowercase_ = """SECRET""" lowercase_ = generate_key(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase_ = cipher_text(lowerCAmelCase__ , lowerCAmelCase__ ) print(F'''Encrypted Text = {s}''' ) print(F'''Original Text = {original_text(lowerCAmelCase__ , lowerCAmelCase__ )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
361
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def _SCREAMING_SNAKE_CASE () -> Generator[int, None, None]: '''simple docstring''' lowercase_ = {} lowercase_ = 2 while True: lowercase_ = factor_map.pop(__lowerCAmelCase , __lowerCAmelCase ) if factor: lowercase_ = factor + prime while x in factor_map: x += factor lowercase_ = factor else: lowercase_ = prime yield prime prime += 1 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase = 1E10 ) -> int: '''simple docstring''' lowercase_ = sieve() lowercase_ = 1 while True: lowercase_ = next(__lowerCAmelCase ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(__lowerCAmelCase ) n += 2 if __name__ == "__main__": print(solution())
313
0
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase : str = '''▁''' lowerCAmelCase : Dict = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class _A ( __magic_name__ , unittest.TestCase): SCREAMING_SNAKE_CASE : Tuple = BigBirdTokenizer SCREAMING_SNAKE_CASE : Dict = BigBirdTokenizerFast SCREAMING_SNAKE_CASE : List[str] = True SCREAMING_SNAKE_CASE : List[Any] = True def UpperCAmelCase ( self ): """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE_ : int = self.tokenizer_class(UpperCAmelCase_ , keep_accents=UpperCAmelCase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = '<s>' SCREAMING_SNAKE_CASE_ : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_ ) , UpperCAmelCase_ ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '[MASK]' ) self.assertEqual(len(UpperCAmelCase_ ) , 1004 ) def UpperCAmelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCAmelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE_ : int = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.tokenize(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : str = rust_tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : str = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.encode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : Any = rust_tokenizer.encode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = BigBirdTokenizer(UpperCAmelCase_ , keep_accents=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : int = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase_ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [285, 46, 10, 170, 382] , ) SCREAMING_SNAKE_CASE_ : Tuple = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) self.assertListEqual( UpperCAmelCase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual( UpperCAmelCase_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase ( self ): """simple docstring""" return BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = 'Hello World!' SCREAMING_SNAKE_CASE_ : Optional[int] = [65, 1_8536, 2260, 101, 66] self.assertListEqual(UpperCAmelCase_ , self.big_tokenizer.encode(UpperCAmelCase_ ) ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) # fmt: off SCREAMING_SNAKE_CASE_ : Optional[int] = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 3_4324, 497, 391, 408, 1_1342, 1244, 385, 100, 938, 985, 456, 574, 362, 1_2597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(UpperCAmelCase_ , self.big_tokenizer.encode(UpperCAmelCase_ ) ) @require_torch @slow def UpperCAmelCase ( self ): """simple docstring""" import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence SCREAMING_SNAKE_CASE_ : Optional[int] = list(self.big_tokenizer.get_vocab().keys() )[:10] SCREAMING_SNAKE_CASE_ : Optional[int] = ' '.join(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : int = self.big_tokenizer.encode_plus(UpperCAmelCase_ , return_tensors='pt' , return_token_type_ids=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : List[str] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE_ : List[Any] = BigBirdConfig(attention_type='original_full' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = BigBirdModel(UpperCAmelCase_ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase_ ) model(**UpperCAmelCase_ ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.decode(tokenizer('Paris is the [MASK].' ).input_ids ) self.assertTrue(decoded_text == '[CLS] Paris is the[MASK].[SEP]' ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = {'input_ids': [[65, 3_9286, 458, 3_6335, 2001, 456, 1_3073, 1_3266, 455, 113, 7746, 1741, 1_1157, 391, 1_3073, 1_3266, 455, 113, 3967, 3_5412, 113, 4936, 109, 3870, 2377, 113, 3_0084, 4_5720, 458, 134, 1_7496, 112, 503, 1_1672, 113, 118, 112, 5665, 1_3347, 3_8687, 112, 1496, 3_1389, 112, 3268, 4_7264, 134, 962, 112, 1_6377, 8035, 2_3130, 430, 1_2169, 1_5518, 2_8592, 458, 146, 4_1697, 109, 391, 1_2169, 1_5518, 1_6689, 458, 146, 4_1358, 109, 452, 726, 4034, 111, 763, 3_5412, 5082, 388, 1903, 111, 9051, 391, 2870, 4_8918, 1900, 1123, 550, 998, 112, 9586, 1_5985, 455, 391, 410, 2_2955, 3_7636, 114, 66], [65, 448, 1_7496, 419, 3663, 385, 763, 113, 2_7533, 2870, 3283, 1_3043, 1639, 2_4713, 523, 656, 2_4013, 1_8550, 2521, 517, 2_7014, 2_1244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 1_1786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 2_1932, 1_8146, 726, 363, 1_7032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ , model_name='google/bigbird-roberta-base' , revision='215c99f1600e06f83acce68422f2035b2b5c3510' , )
253
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> str: '''simple docstring''' return " ".join( ''.join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
138
0
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Dict = 1.5 A_ : Union[str, Any] = int(factor * num_class_images ) A_ : List[Any] = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=lowerCamelCase__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=lowerCamelCase__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: A_ : str = client.query(text=lowerCamelCase__ ) if len(lowerCamelCase__ ) >= factor * num_class_images or num_images > 1E4: break else: A_ : Optional[Any] = int(factor * num_images ) A_ : int = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=lowerCamelCase__ , aesthetic_weight=0.1 , ) A_ : Tuple = 0 A_ : List[Any] = 0 A_ : Dict = tqdm(desc="""downloading real regularization images""" , total=lowerCamelCase__ ) with open(f'{class_data_dir}/caption.txt' , """w""" ) as fa, open(f'{class_data_dir}/urls.txt' , """w""" ) as fa, open( f'{class_data_dir}/images.txt' , """w""" ) as fa: while total < num_class_images: A_ : Union[str, Any] = class_images[count] count += 1 try: A_ : Optional[int] = requests.get(images["""url"""] ) if img.status_code == 2_00: A_ : List[Any] = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , """wb""" ) as f: f.write(img.content ) fa.write(images["""caption"""] + """\n""" ) fa.write(images["""url"""] + """\n""" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + """\n""" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def a ( ): '''simple docstring''' A_ : Union[str, Any] = argparse.ArgumentParser("""""" , add_help=lowerCamelCase__ ) parser.add_argument("""--class_prompt""" , help="""text prompt to retrieve images""" , required=lowerCamelCase__ , type=lowerCamelCase__ ) parser.add_argument("""--class_data_dir""" , help="""path to save images""" , required=lowerCamelCase__ , type=lowerCamelCase__ ) parser.add_argument("""--num_class_images""" , help="""number of images to download""" , default=2_00 , type=lowerCamelCase__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase :Tuple = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
135
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase :Any = logging.get_logger(__name__) lowerCamelCase :List[Any] = { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/config.json''', '''umberto-commoncrawl-cased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json''' ), '''umberto-wikipedia-uncased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json''' ), } class _lowerCAmelCase ( __UpperCAmelCase ): __SCREAMING_SNAKE_CASE : Optional[int] = 'camembert' def __init__(self , lowercase=30522 , lowercase=768 , lowercase=12 , lowercase=12 , lowercase=3072 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=2 , lowercase=0.02 , lowercase=1E-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ): super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase ) A_ : List[Any] = vocab_size A_ : int = hidden_size A_ : Dict = num_hidden_layers A_ : Dict = num_attention_heads A_ : Optional[Any] = hidden_act A_ : str = intermediate_size A_ : int = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Optional[Any] = max_position_embeddings A_ : Optional[int] = type_vocab_size A_ : int = initializer_range A_ : str = layer_norm_eps A_ : int = position_embedding_type A_ : Dict = use_cache A_ : Any = classifier_dropout class _lowerCAmelCase ( __UpperCAmelCase ): @property def _a (self ): if self.task == "multiple-choice": A_ : Optional[Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A_ : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
135
1
def __lowercase ( a__ ) -> bool: return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('''Program to check whether a number is a Perfect number or not...''') lowerCAmelCase__ : Optional[Any] =int(input('''Enter number: ''').strip()) print(F'''{number} is {'' if perfect(number) else 'not '}a Perfect Number.''')
257
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCAmelCase__ : List[Any] =input('''Enter image url: ''').strip() print(F'''Downloading image from {url} ...''') lowerCAmelCase__ : int =BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCAmelCase__ : Union[str, Any] =soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCAmelCase__ : int =requests.get(image_url).content lowerCAmelCase__ : Optional[int] =F'''{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg''' with open(file_name, '''wb''') as fp: fp.write(image_data) print(F'''Done. Image saved to disk as {file_name}.''')
257
1
def lowerCAmelCase_ ( __lowerCAmelCase )-> list[list]: '''simple docstring''' UpperCAmelCase : int =current_set.copy() for row_index, row in enumerate(__lowerCAmelCase ): UpperCAmelCase : Optional[Any] =row[0] for column_index, column in enumerate(__lowerCAmelCase ): if magnitude == 0: UpperCAmelCase : Optional[int] =column continue UpperCAmelCase : str =column / magnitude # Subtract to cancel term UpperCAmelCase : Optional[Any] =current_set[0] UpperCAmelCase : Any =[first_row] UpperCAmelCase : Optional[Any] =current_set[1::] for row in current_set: UpperCAmelCase : Tuple =[] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(__lowerCAmelCase ) continue for column_index in range(len(__lowerCAmelCase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(__lowerCAmelCase ) # Create next recursion iteration set if len(final_set[0] ) != 3: UpperCAmelCase : Any =final_set[0] UpperCAmelCase : str =[] UpperCAmelCase : str =[] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) UpperCAmelCase : int =simplify(__lowerCAmelCase ) for i in range(len(__lowerCAmelCase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , __lowerCAmelCase ) UpperCAmelCase : Dict =resultant return final_set def lowerCAmelCase_ ( __lowerCAmelCase )-> list: '''simple docstring''' if len(__lowerCAmelCase ) == 0: raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) UpperCAmelCase : Tuple =len(__lowerCAmelCase ) + 1 if any(len(__lowerCAmelCase ) != _length for item in equations ): raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) for row in equations: if any(not isinstance(__lowerCAmelCase , (int, float) ) for column in row ): raise ValueError('''solve_simultaneous() requires lists of integers''' ) if len(__lowerCAmelCase ) == 1: return [equations[0][-1] / equations[0][0]] UpperCAmelCase : Tuple =equations.copy() if any(0 in row for row in data_set ): UpperCAmelCase : int =data_set.copy() UpperCAmelCase : List[Any] =[] for row_index, row in enumerate(__lowerCAmelCase ): if 0 not in row: UpperCAmelCase : str =data_set.pop(__lowerCAmelCase ) break if not full_row: raise ValueError('''solve_simultaneous() requires at least 1 full equation''' ) data_set.insert(0 , __lowerCAmelCase ) UpperCAmelCase : Union[str, Any] =data_set.copy() UpperCAmelCase : Optional[int] =simplify(__lowerCAmelCase ) UpperCAmelCase : List[str] =simplified[::-1] UpperCAmelCase : list =[] for row in simplified: UpperCAmelCase : List[Any] =row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue UpperCAmelCase : Any =row.copy()[: len(__lowerCAmelCase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(__lowerCAmelCase ) == 0: solutions.append(0 ) continue UpperCAmelCase : Any =temp_row[1::] UpperCAmelCase : Optional[Any] =temp_row[::-1] for column_index, column in enumerate(__lowerCAmelCase ): current_solution -= column * solutions[column_index] solutions.append(__lowerCAmelCase ) UpperCAmelCase : int =[] for item in solutions: final.append(float(round(__lowerCAmelCase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() __snake_case = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
78
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __snake_case ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase : Dict = StableUnCLIPPipeline __lowerCamelCase : int = TEXT_TO_IMAGE_PARAMS __lowerCamelCase : int = TEXT_TO_IMAGE_BATCH_PARAMS __lowerCamelCase : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS __lowerCamelCase : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false __lowerCamelCase : Optional[Any] = False def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : int =32 UpperCAmelCase : Union[str, Any] =embedder_hidden_size # prior components torch.manual_seed(0 ) UpperCAmelCase : Optional[Any] =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) UpperCAmelCase : int =CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case__ , projection_dim=snake_case__ , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) UpperCAmelCase : Dict =PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=snake_case__ , num_layers=1 , ) torch.manual_seed(0 ) UpperCAmelCase : Tuple =DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=snake_case__ , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) UpperCAmelCase : Optional[int] =StableUnCLIPImageNormalizer(embedding_dim=snake_case__ ) UpperCAmelCase : Any =DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) UpperCAmelCase : List[str] =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) UpperCAmelCase : List[str] =CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case__ , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) UpperCAmelCase : Optional[int] =UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=snake_case__ , layers_per_block=1 , upcast_attention=snake_case__ , use_linear_projection=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase : List[Any] =DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_0085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=snake_case__ , steps_offset=1 , ) torch.manual_seed(0 ) UpperCAmelCase : Dict =AutoencoderKL() UpperCAmelCase : Tuple ={ # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def UpperCAmelCase__ ( self , snake_case__ , snake_case__=0 ) -> List[Any]: '''simple docstring''' if str(snake_case__ ).startswith('''mps''' ): UpperCAmelCase : Union[str, Any] =torch.manual_seed(snake_case__ ) else: UpperCAmelCase : Any =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase : str ={ '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Tuple =torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=snake_case__ ) def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' UpperCAmelCase : List[Any] =torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=snake_case__ ) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) UpperCAmelCase : Optional[int] =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() UpperCAmelCase : int =torch.Generator(device='''cpu''' ).manual_seed(0 ) UpperCAmelCase : int =pipe('''anime turle''' , generator=snake_case__ , output_type='''np''' ) UpperCAmelCase : str =output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ ) def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCAmelCase : List[str] =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) UpperCAmelCase : str =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() UpperCAmelCase : Any =pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) UpperCAmelCase : Tuple =torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
78
1
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class a__ ( snake_case__ ): """simple docstring""" def __init__( self : str , *UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Any=None , **UpperCAmelCase__ : Optional[Any] ) ->Tuple: """simple docstring""" super().__init__(*snake_case__ , **snake_case__ ) SCREAMING_SNAKE_CASE : List[Any] = eval_examples SCREAMING_SNAKE_CASE : Dict = post_process_function def _lowercase ( self : int , UpperCAmelCase__ : Optional[Dataset] = None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Optional[List[str]] = None , UpperCAmelCase__ : str = "eval" , **UpperCAmelCase__ : Any , ) ->str: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = gen_kwargs.copy() SCREAMING_SNAKE_CASE : List[str] = ( gen_kwargs["max_length"] if gen_kwargs.get("""max_length""" ) is not None else self.args.generation_max_length ) SCREAMING_SNAKE_CASE : Dict = ( gen_kwargs["num_beams"] if gen_kwargs.get("""num_beams""" ) is not None else self.args.generation_num_beams ) SCREAMING_SNAKE_CASE : Optional[int] = gen_kwargs SCREAMING_SNAKE_CASE : Any = self.eval_dataset if eval_dataset is None else eval_dataset SCREAMING_SNAKE_CASE : str = self.get_eval_dataloader(snake_case__ ) SCREAMING_SNAKE_CASE : Optional[int] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : List[Any] = self.compute_metrics SCREAMING_SNAKE_CASE : List[Any] = None SCREAMING_SNAKE_CASE : Dict = time.time() SCREAMING_SNAKE_CASE : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : Optional[Any] = eval_loop( snake_case__ , description="""Evaluation""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=snake_case__ , metric_key_prefix=snake_case__ , ) finally: SCREAMING_SNAKE_CASE : Optional[int] = compute_metrics SCREAMING_SNAKE_CASE : Tuple = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( snake_case__ , snake_case__ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default SCREAMING_SNAKE_CASE : Union[str, Any] = self.post_process_function(snake_case__ , snake_case__ , snake_case__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = self.compute_metrics(snake_case__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): SCREAMING_SNAKE_CASE : List[Any] = metrics.pop(snake_case__ ) metrics.update(output.metrics ) else: SCREAMING_SNAKE_CASE : Dict = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(snake_case__ ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) SCREAMING_SNAKE_CASE : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , snake_case__ ) return metrics def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : str = "test" , **UpperCAmelCase__ : Tuple ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = gen_kwargs.copy() SCREAMING_SNAKE_CASE : List[Any] = self.get_test_dataloader(snake_case__ ) # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : Any = self.compute_metrics SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : List[str] = time.time() SCREAMING_SNAKE_CASE : Tuple = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : Optional[int] = eval_loop( snake_case__ , description="""Prediction""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=snake_case__ , metric_key_prefix=snake_case__ , ) finally: SCREAMING_SNAKE_CASE : Union[str, Any] = compute_metrics SCREAMING_SNAKE_CASE : Union[str, Any] = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( snake_case__ , snake_case__ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output SCREAMING_SNAKE_CASE : int = self.post_process_function(snake_case__ , snake_case__ , snake_case__ , """predict""" ) SCREAMING_SNAKE_CASE : Optional[int] = self.compute_metrics(snake_case__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): SCREAMING_SNAKE_CASE : Tuple = metrics.pop(snake_case__ ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=snake_case__ )
245
"""simple docstring""" import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin A__ : Union[str, Any] = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class lowercase__ : def __init__( self : List[Any] , snake_case__ : int , snake_case__ : List[str]=16 , snake_case__ : Tuple=13 , snake_case__ : Dict=7 , snake_case__ : List[Any]=14 , snake_case__ : List[Any]=10 , snake_case__ : Dict=19 , snake_case__ : List[str]=5 , snake_case__ : Union[str, Any]=4 , snake_case__ : str=True , snake_case__ : int=16 , snake_case__ : Union[str, Any]=2 , snake_case__ : Tuple=4 , snake_case__ : Dict=4 , snake_case__ : int="gelu" , snake_case__ : Dict=0.1 , snake_case__ : str=0.1 , snake_case__ : List[str]=[1, 2, 3, 4, 5] , snake_case__ : Optional[int]=25 , snake_case__ : Dict=5 , ): lowerCamelCase_ : Dict =d_model lowerCamelCase_ : int =parent lowerCamelCase_ : Optional[Any] =batch_size lowerCamelCase_ : int =prediction_length lowerCamelCase_ : Optional[int] =context_length lowerCamelCase_ : Any =cardinality lowerCamelCase_ : List[str] =num_time_features lowerCamelCase_ : List[Any] =lags_sequence lowerCamelCase_ : Optional[int] =embedding_dimension lowerCamelCase_ : Union[str, Any] =is_training lowerCamelCase_ : Union[str, Any] =hidden_size lowerCamelCase_ : str =num_hidden_layers lowerCamelCase_ : Any =num_attention_heads lowerCamelCase_ : Any =intermediate_size lowerCamelCase_ : Union[str, Any] =hidden_act lowerCamelCase_ : Optional[int] =hidden_dropout_prob lowerCamelCase_ : Optional[int] =attention_probs_dropout_prob lowerCamelCase_ : List[Any] =context_length lowerCamelCase_ : str =prediction_length + label_length lowerCamelCase_ : int =label_length lowerCamelCase_ : Union[str, Any] =moving_average lowerCamelCase_ : str =autocorrelation_factor def UpperCAmelCase__ ( self : Any ): return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : List[Any] ): lowerCamelCase_ : Optional[Any] =config.context_length + max(config.lags_sequence ) lowerCamelCase_ : Any =ids_tensor([self.batch_size, 1] , config.cardinality[0] ) lowerCamelCase_ : List[Any] =floats_tensor([self.batch_size, _past_length, config.num_time_features] ) lowerCamelCase_ : List[str] =floats_tensor([self.batch_size, _past_length] ) lowerCamelCase_ : Any =floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs lowerCamelCase_ : Tuple =floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) lowerCamelCase_ : Optional[Any] =floats_tensor([self.batch_size, config.prediction_length] ) lowerCamelCase_ : Any ={ "past_values": past_values, "static_categorical_features": static_categorical_features, "past_time_features": past_time_features, "past_observed_mask": past_observed_mask, "future_time_features": future_time_features, "future_values": future_values, } return inputs_dict def UpperCAmelCase__ ( self : Tuple ): lowerCamelCase_ : str =self.get_config() lowerCamelCase_ : List[Any] =self.prepare_autoformer_inputs_dict(snake_case__ ) return config, inputs_dict def UpperCAmelCase__ ( self : str ): lowerCamelCase_ , lowerCamelCase_ : List[str] =self.prepare_config_and_inputs() return config, inputs_dict def UpperCAmelCase__ ( self : List[Any] , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] ): lowerCamelCase_ : str =AutoformerModel(config=snake_case__ ).to(snake_case__ ).eval() lowerCamelCase_ : int =model(**snake_case__ ) lowerCamelCase_ : str =outputs.encoder_last_hidden_state lowerCamelCase_ : Optional[Any] =outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ : Tuple =model.get_encoder() encoder.save_pretrained(snake_case__ ) lowerCamelCase_ : Any =AutoformerEncoder.from_pretrained(snake_case__ ).to(snake_case__ ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Optional[Any] =model.create_network_inputs(**snake_case__ ) lowerCamelCase_ , lowerCamelCase_ : Optional[int] =model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) lowerCamelCase_ : Dict =torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) lowerCamelCase_ : int =encoder(inputs_embeds=snake_case__ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) lowerCamelCase_ : str =( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) lowerCamelCase_ : Optional[int] =torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) lowerCamelCase_ : Any =torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) lowerCamelCase_ : Optional[Any] =torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ : List[str] =model.get_decoder() decoder.save_pretrained(snake_case__ ) lowerCamelCase_ : str =AutoformerDecoder.from_pretrained(snake_case__ ).to(snake_case__ ) lowerCamelCase_ : List[str] =decoder( trend=snake_case__ , inputs_embeds=snake_case__ , encoder_hidden_states=snake_case__ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class lowercase__ ( snake_case__, snake_case__, unittest.TestCase ): _UpperCAmelCase :Optional[int] = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () _UpperCAmelCase :Union[str, Any] = (AutoformerForPrediction,) if is_torch_available() else () _UpperCAmelCase :Optional[int] = {"feature-extraction": AutoformerModel} if is_torch_available() else {} _UpperCAmelCase :Tuple = False _UpperCAmelCase :int = False _UpperCAmelCase :int = False _UpperCAmelCase :Optional[int] = False _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :Dict = False def UpperCAmelCase__ ( self : Optional[int] ): lowerCamelCase_ : List[str] =AutoformerModelTester(self ) lowerCamelCase_ : List[str] =ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def UpperCAmelCase__ ( self : Tuple ): self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : Union[str, Any] ): lowerCamelCase_ , lowerCamelCase_ : str =self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: lowerCamelCase_ : List[Any] =model_class(snake_case__ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case__ ) lowerCamelCase_ , lowerCamelCase_ : str =model_class.from_pretrained(snake_case__ , output_loading_info=snake_case__ ) self.assertEqual(info["missing_keys"] , [] ) def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*snake_case__ ) @unittest.skip(reason="Model has no tokens embeddings" ) def UpperCAmelCase__ ( self : Optional[Any] ): pass def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : Any =inspect.signature(getattr(snake_case__ , "forward" ) ) # The main input is the name of the argument after `self` lowerCamelCase_ : Optional[Any] =list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , snake_case__ ) def UpperCAmelCase__ ( self : List[str] ): lowerCamelCase_ , lowerCamelCase_ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ : Optional[int] =model_class(snake_case__ ) lowerCamelCase_ : Optional[int] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ : Union[str, Any] =[*signature.parameters.keys()] lowerCamelCase_ : List[Any] =[ "past_values", "past_time_features", "past_observed_mask", "static_categorical_features", "static_real_features", "future_values", "future_time_features", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("future_observed_mask" ) expected_arg_names.extend( [ "decoder_attention_mask", "head_mask", "decoder_head_mask", "cross_attn_head_mask", "encoder_outputs", "past_key_values", "output_hidden_states", "output_attentions", "use_cache", "return_dict", ] ) self.assertListEqual(arg_names[: len(snake_case__ )] , snake_case__ ) def UpperCAmelCase__ ( self : Tuple ): lowerCamelCase_ , lowerCamelCase_ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ : Optional[int] =True lowerCamelCase_ : List[str] =getattr(self.model_tester , "seq_length" , snake_case__ ) lowerCamelCase_ : Dict =getattr(self.model_tester , "decoder_seq_length" , snake_case__ ) lowerCamelCase_ : List[Any] =getattr(self.model_tester , "encoder_seq_length" , snake_case__ ) lowerCamelCase_ : Optional[Any] =getattr(self.model_tester , "d_model" , snake_case__ ) lowerCamelCase_ : List[str] =getattr(self.model_tester , "num_attention_heads" , snake_case__ ) lowerCamelCase_ : Union[str, Any] =d_model // num_attention_heads for model_class in self.all_model_classes: lowerCamelCase_ : str =True lowerCamelCase_ : int =False lowerCamelCase_ : Any =True lowerCamelCase_ : Tuple =model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): lowerCamelCase_ : Union[str, Any] =model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) lowerCamelCase_ : str =outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCamelCase_ : List[Any] =True lowerCamelCase_ : Optional[int] =model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): lowerCamelCase_ : List[str] =model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) lowerCamelCase_ : Union[str, Any] =outputs.encoder_attentions self.assertEqual(len(snake_case__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) lowerCamelCase_ : Optional[Any] =len(snake_case__ ) lowerCamelCase_ : List[Any] =7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(snake_case__ , snake_case__ ) # decoder attentions lowerCamelCase_ : Union[str, Any] =outputs.decoder_attentions self.assertIsInstance(snake_case__ , (list, tuple) ) self.assertEqual(len(snake_case__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions lowerCamelCase_ : Tuple =outputs.cross_attentions self.assertIsInstance(snake_case__ , (list, tuple) ) self.assertEqual(len(snake_case__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine lowerCamelCase_ : Tuple =True lowerCamelCase_ : Optional[int] =True lowerCamelCase_ : Tuple =model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): lowerCamelCase_ : Dict =model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) self.assertEqual(out_len + 2 , len(snake_case__ ) ) lowerCamelCase_ : Union[str, Any] =outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def UpperCAmelCase__ ( self : Optional[int] ): super().test_retain_grad_hidden_states_attentions() def _snake_case ( lowerCamelCase__ : Tuple="train-batch.pt" ) -> Any: lowerCamelCase_ : Tuple =hf_hub_download(repo_id="hf-internal-testing/tourism-monthly-batch" , filename=lowerCamelCase__ , repo_type="dataset" ) lowerCamelCase_ : List[Any] =torch.load(lowerCamelCase__ , map_location=lowerCamelCase__ ) return batch @require_torch @slow class lowercase__ ( unittest.TestCase ): def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : int =AutoformerModel.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(snake_case__ ) lowerCamelCase_ : List[str] =prepare_batch() with torch.no_grad(): lowerCamelCase_ : List[Any] =model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , future_values=batch["future_values"] , future_time_features=batch["future_time_features"] , )[0] lowerCamelCase_ : Union[str, Any] =torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , snake_case__ ) lowerCamelCase_ : Dict =torch.tensor( [[0.3_593, -1.3_398, 0.6_330], [0.2_279, 1.5_396, -0.1_792], [0.0_450, 1.3_225, -0.2_335]] , device=snake_case__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case__ , atol=snake_case__ ) ) def UpperCAmelCase__ ( self : Tuple ): lowerCamelCase_ : str =AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(snake_case__ ) lowerCamelCase_ : Optional[int] =prepare_batch("val-batch.pt" ) with torch.no_grad(): lowerCamelCase_ : Union[str, Any] =model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , ).encoder_last_hidden_state lowerCamelCase_ : List[Any] =torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , snake_case__ ) lowerCamelCase_ : Optional[Any] =torch.tensor( [[-0.0_734, -0.9_036, 0.8_358], [4.7_186, 2.4_113, 1.9_581], [1.7_953, 2.3_558, 1.2_970]] , device=snake_case__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case__ , atol=snake_case__ ) ) def UpperCAmelCase__ ( self : List[str] ): lowerCamelCase_ : int =AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(snake_case__ ) lowerCamelCase_ : Dict =prepare_batch("val-batch.pt" ) with torch.no_grad(): lowerCamelCase_ : Union[str, Any] =model.generate( static_categorical_features=batch["static_categorical_features"] , past_time_features=batch["past_time_features"] , past_values=batch["past_values"] , future_time_features=batch["future_time_features"] , past_observed_mask=batch["past_observed_mask"] , ) lowerCamelCase_ : Tuple =torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , snake_case__ ) lowerCamelCase_ : List[str] =torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=snake_case__ ) lowerCamelCase_ : Any =outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , snake_case__ , rtol=1E-1 ) )
144
0
"""simple docstring""" from __future__ import annotations from fractions import Fraction def __A ( a_ :int , a_ :int) -> bool: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __A ( a_ :int) -> list[str]: __a : Union[str, Any] = [] __a : int = 11 __a : Optional[Any] = int('''1''' + '''0''' * digit_len) for num in range(a_ , a_): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(a_ , a_): solutions.append(F"""{num}/{den}""") den += 1 num += 1 __a : Tuple = 10 return solutions def __A ( a_ :int = 2) -> int: __a : Any = 1.0 for fraction in fraction_list(a_): __a : int = Fraction(a_) result *= frac.denominator / frac.numerator return int(a_) if __name__ == "__main__": print(solution())
188
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets A = datasets.logging.get_logger(__name__) A = '''\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = "{COMET}: A Neural Framework for {MT} Evaluation", author = "Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.213", pages = "2685--2702", } ''' A = '''\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. ''' A = ''' COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric(\'comet\') >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use >>> source = ["Dem Feuer konnte Einhalt geboten werden", "Schulen und Kindergärten wurden eröffnet."] >>> hypothesis = ["The fire could be stopped", "Schools and kindergartens were open"] >>> reference = ["They were able to control the fire.", "Schools and kindergartens opened"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results["scores"]]) [0.19, 0.92] ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): '''simple docstring''' def _lowerCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://unbabel.github.io/COMET/html/index.html''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''sources''': datasets.Value('''string''' , id='''sequence''' ), '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/Unbabel/COMET'''] , reference_urls=[ '''https://github.com/Unbabel/COMET''', '''https://www.aclweb.org/anthology/2020.emnlp-main.213/''', '''http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6''', ] , ) def _lowerCamelCase ( self , _UpperCAmelCase ): if self.config_name == "default": __a : List[str] = comet.load_from_checkpoint(comet.download_model('''wmt20-comet-da''' ) ) else: __a : List[str] = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=False ): if gpus is None: __a : str = 1 if torch.cuda.is_available() else 0 __a : Optional[Any] = {'''src''': sources, '''mt''': predictions, '''ref''': references} __a : Dict = [dict(zip(_UpperCAmelCase , _UpperCAmelCase ) ) for t in zip(*data.values() )] __a , __a : int = self.scorer.predict(_UpperCAmelCase , gpus=_UpperCAmelCase , progress_bar=_UpperCAmelCase ) return {"mean_score": mean_score, "scores": scores}
188
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : Dict = {"configuration_vit_msn": ["VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMSNConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = [ "VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMSNModel", "ViTMSNForImageClassification", "ViTMSNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys __lowerCAmelCase : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
156
from collections import deque from .hash_table import HashTable class __lowerCAmelCase ( lowerCAmelCase_ ): """simple docstring""" def __init__( self : Union[str, Any] , *_snake_case : Union[str, Any] , **_snake_case : Union[str, Any] ): super().__init__(*_snake_case , **_snake_case ) def snake_case_ ( self : List[Any] , _snake_case : List[Any] , _snake_case : Dict ): __lowercase : Any = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_snake_case ) __lowercase : List[Any] = self.values[key] def snake_case_ ( self : Any ): return ( sum(self.charge_factor - len(_snake_case ) for slot in self.values ) / self.size_table * self.charge_factor ) def snake_case_ ( self : int , _snake_case : str , _snake_case : Optional[int]=None ): if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_snake_case ) == 0 ): return key return super()._collision_resolution(_snake_case , _snake_case )
156
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> List[Any]: '''simple docstring''' return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(lowerCamelCase__ ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__("doctest").testmod()
350
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html UpperCAmelCase : Optional[Any] = "platform" import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class SCREAMING_SNAKE_CASE__ : lowercase__ = PegasusConfig lowercase__ = {} lowercase__ = "gelu" def __init__( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=1_3 , lowerCAmelCase_ : Any=7 , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : str=9_9 , lowerCAmelCase_ : Tuple=3_2 , lowerCAmelCase_ : Dict=5 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Dict=3_7 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Optional[int]=2_0 , lowerCAmelCase_ : Tuple=2 , lowerCAmelCase_ : List[str]=1 , lowerCAmelCase_ : Optional[Any]=0 , ): """simple docstring""" lowercase_ = parent lowercase_ = batch_size lowercase_ = seq_length lowercase_ = is_training lowercase_ = use_labels lowercase_ = vocab_size lowercase_ = hidden_size lowercase_ = num_hidden_layers lowercase_ = num_attention_heads lowercase_ = intermediate_size lowercase_ = hidden_dropout_prob lowercase_ = attention_probs_dropout_prob lowercase_ = max_position_embeddings lowercase_ = eos_token_id lowercase_ = pad_token_id lowercase_ = bos_token_id def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" lowercase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size).clip(3 , self.vocab_size) lowercase_ = np.expand_dims(np.array([self.eos_token_id] * self.batch_size) , 1) lowercase_ = np.concatenate([input_ids, eos_tensor] , axis=1) lowercase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowercase_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowercase_ = prepare_pegasus_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) return config, inputs_dict def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Optional[Any]): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""") lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1])), ] , axis=-1 , ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowercase_ = np.not_equal(__lowerCAmelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: lowercase_ = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , unittest.TestCase ): lowercase__ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) lowercase__ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ = FlaxPegasusModelTester(self) lowercase_ = ConfigTester(self , config_class=lowerCAmelCase_) def _UpperCAmelCase ( self : Any): """simple docstring""" self.config_tester.run_common_tests() def _UpperCAmelCase ( self : List[str]): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model_class(lowerCAmelCase_) @jax.jit def encode_jitted(lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[int]=None , **lowerCAmelCase_ : Optional[int]): return model.encode(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_) with self.subTest("""JIT Enabled"""): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = model_class(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""]) lowercase_ = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict): return model.decode( decoder_input_ids=lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , encoder_outputs=lowerCAmelCase_ , ) with self.subTest("""JIT Enabled"""): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) @slow def _UpperCAmelCase ( self : Tuple): """simple docstring""" for model_class_name in self.all_model_classes: lowercase_ = model_class_name.from_pretrained("""google/pegasus-large""" , from_pt=lowerCAmelCase_) lowercase_ = np.ones((1, 1)) lowercase_ = model(lowerCAmelCase_) self.assertIsNotNone(lowerCAmelCase_) @slow def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""") lowercase_ = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""") lowercase_ = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] lowercase_ = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] lowercase_ = tokenizer(lowerCAmelCase_ , return_tensors="""np""" , truncation=lowerCAmelCase_ , max_length=5_1_2 , padding=lowerCAmelCase_) lowercase_ = model.generate(**lowerCAmelCase_ , num_beams=2).sequences lowercase_ = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_) assert tgt_text == decoded
313
0
'''simple docstring''' def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _a , _a : int = len(lowerCAmelCase_ ), len(grid[0] ) if ( min(lowerCAmelCase_ , lowerCAmelCase_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) _a : int = 0 count += depth_first_search(lowerCAmelCase_ , row + 1 , lowerCAmelCase_ , lowerCAmelCase_ ) count += depth_first_search(lowerCAmelCase_ , row - 1 , lowerCAmelCase_ , lowerCAmelCase_ ) count += depth_first_search(lowerCAmelCase_ , lowerCAmelCase_ , col + 1 , lowerCAmelCase_ ) count += depth_first_search(lowerCAmelCase_ , lowerCAmelCase_ , col - 1 , lowerCAmelCase_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
89
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self : Dict , lowerCamelCase_ : Dict[str, int] , lowerCamelCase_ : List[str] , lowerCamelCase_ : int = None , lowerCamelCase_ : int = None ): '''simple docstring''' super().__init__() SCREAMING_SNAKE_CASE : Any = pad_token_id SCREAMING_SNAKE_CASE : List[Any] = max_length SCREAMING_SNAKE_CASE : Optional[int] = vocab SCREAMING_SNAKE_CASE : List[Any] = merges SCREAMING_SNAKE_CASE : Tuple = BytePairTokenizer(lowerCamelCase_ , lowerCamelCase_ , sequence_length=lowerCamelCase_ ) @classmethod def lowerCamelCase_ ( cls : Any , lowerCamelCase_ : GPTaTokenizer , *lowerCamelCase_ : str , **lowerCamelCase_ : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = [""" """.join(lowerCamelCase_ ) for m in tokenizer.bpe_ranks.keys()] SCREAMING_SNAKE_CASE : List[str] = tokenizer.get_vocab() return cls(lowerCamelCase_ , lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) @classmethod def lowerCamelCase_ ( cls : List[Any] , lowerCamelCase_ : Union[str, os.PathLike] , *lowerCamelCase_ : str , **lowerCamelCase_ : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = GPTaTokenizer.from_pretrained(lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) return cls.from_tokenizer(lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) @classmethod def lowerCamelCase_ ( cls : List[str] , lowerCamelCase_ : Tuple ): '''simple docstring''' return cls(**lowerCamelCase_ ) def lowerCamelCase_ ( self : Dict ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : Dict , lowerCamelCase_ : int = None ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = self.tf_tokenizer(lowerCamelCase_ ) SCREAMING_SNAKE_CASE : Optional[int] = tf.ones_like(lowerCamelCase_ ) if self.pad_token_id is not None: # pad the tokens up to max length SCREAMING_SNAKE_CASE : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE : Union[str, Any] = pad_model_inputs( lowerCamelCase_ , max_seq_length=lowerCamelCase_ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
323
0
"""simple docstring""" import numpy as np def UpperCAmelCase__ ( lowerCAmelCase__ :np.array ) -> np.array: '''simple docstring''' return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
32
"""simple docstring""" import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : List[Any] =logging.get_logger(__name__) def UpperCAmelCase__ ( lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] ) -> int: '''simple docstring''' lowercase = [] for i in range(encoder_config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'encoder.deit.blocks.{i}.norm1.weight', f'encoder.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'encoder.deit.blocks.{i}.norm1.bias', f'encoder.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (f'encoder.deit.blocks.{i}.attn.proj.weight', f'encoder.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append( (f'encoder.deit.blocks.{i}.attn.proj.bias', f'encoder.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append( (f'encoder.deit.blocks.{i}.norm2.weight', f'encoder.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'encoder.deit.blocks.{i}.norm2.bias', f'encoder.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append( (f'encoder.deit.blocks.{i}.mlp.fc1.weight', f'encoder.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append( (f'encoder.deit.blocks.{i}.mlp.fc1.bias', f'encoder.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append( (f'encoder.deit.blocks.{i}.mlp.fc2.weight', f'encoder.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'encoder.deit.blocks.{i}.mlp.fc2.bias', f'encoder.encoder.layer.{i}.output.dense.bias') ) # cls token, position embeddings and patch embeddings of encoder rename_keys.extend( [ ("""encoder.deit.cls_token""", """encoder.embeddings.cls_token"""), ("""encoder.deit.pos_embed""", """encoder.embeddings.position_embeddings"""), ("""encoder.deit.patch_embed.proj.weight""", """encoder.embeddings.patch_embeddings.projection.weight"""), ("""encoder.deit.patch_embed.proj.bias""", """encoder.embeddings.patch_embeddings.projection.bias"""), ("""encoder.deit.norm.weight""", """encoder.layernorm.weight"""), ("""encoder.deit.norm.bias""", """encoder.layernorm.bias"""), ] ) return rename_keys def UpperCAmelCase__ ( lowerCAmelCase__ :str , lowerCAmelCase__ :Any ) -> Dict: '''simple docstring''' for i in range(encoder_config.num_hidden_layers ): # queries, keys and values (only weights, no biases) lowercase = state_dict.pop(f'encoder.deit.blocks.{i}.attn.qkv.weight' ) lowercase = in_proj_weight[ : encoder_config.hidden_size, : ] lowercase = in_proj_weight[ encoder_config.hidden_size : encoder_config.hidden_size * 2, : ] lowercase = in_proj_weight[ -encoder_config.hidden_size :, : ] def UpperCAmelCase__ ( lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :int ) -> Union[str, Any]: '''simple docstring''' lowercase = dct.pop(lowerCAmelCase__ ) lowercase = val def UpperCAmelCase__ ( lowerCAmelCase__ :List[Any] ) -> List[Any]: '''simple docstring''' if "handwritten" in checkpoint_url: lowercase = """https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg""" # industry # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-12.jpg" # have # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-10.jpg" # let # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02.jpg" # # url = "https://fki.tic.heia-fr.ch/static/img/a01-122.jpg" elif "printed" in checkpoint_url or "stage1" in checkpoint_url: lowercase = """https://www.researchgate.net/profile/Dinh-Sang/publication/338099565/figure/fig8/AS:840413229350922@1577381536857/An-receipt-example-in-the-SROIE-2019-dataset_Q640.jpg""" lowercase = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ).convert("""RGB""" ) return im @torch.no_grad() def UpperCAmelCase__ ( lowerCAmelCase__ :int , lowerCAmelCase__ :Union[str, Any] ) -> List[str]: '''simple docstring''' lowercase = ViTConfig(image_size=3_8_4 , qkv_bias=lowerCAmelCase__ ) lowercase = TrOCRConfig() # size of the architecture if "base" in checkpoint_url: lowercase = 7_6_8 elif "large" in checkpoint_url: # use ViT-large encoder lowercase = 1_0_2_4 lowercase = 4_0_9_6 lowercase = 2_4 lowercase = 1_6 lowercase = 1_0_2_4 else: raise ValueError("""Should either find 'base' or 'large' in checkpoint URL""" ) # the large-printed + stage1 checkpoints uses sinusoidal position embeddings, no layernorm afterwards if "large-printed" in checkpoint_url or "stage1" in checkpoint_url: lowercase = False lowercase = """relu""" lowercase = 1_0_2_4 lowercase = True lowercase = False lowercase = False # load HuggingFace model lowercase = ViTModel(lowerCAmelCase__ , add_pooling_layer=lowerCAmelCase__ ) lowercase = TrOCRForCausalLM(lowerCAmelCase__ ) lowercase = VisionEncoderDecoderModel(encoder=lowerCAmelCase__ , decoder=lowerCAmelCase__ ) model.eval() # load state_dict of original model, rename some keys lowercase = torch.hub.load_state_dict_from_url(lowerCAmelCase__ , map_location="""cpu""" , check_hash=lowerCAmelCase__ )["""model"""] lowercase = create_rename_keys(lowerCAmelCase__ , lowerCAmelCase__ ) for src, dest in rename_keys: rename_key(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) read_in_q_k_v(lowerCAmelCase__ , lowerCAmelCase__ ) # remove parameters we don't need del state_dict["encoder.deit.head.weight"] del state_dict["encoder.deit.head.bias"] del state_dict["decoder.version"] # add prefix to decoder keys for key, val in state_dict.copy().items(): lowercase = state_dict.pop(lowerCAmelCase__ ) if key.startswith("""decoder""" ) and "output_projection" not in key: lowercase = val else: lowercase = val # load state dict model.load_state_dict(lowerCAmelCase__ ) # Check outputs on an image lowercase = ViTImageProcessor(size=encoder_config.image_size ) lowercase = RobertaTokenizer.from_pretrained("""roberta-large""" ) lowercase = TrOCRProcessor(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = processor(images=prepare_img(lowerCAmelCase__ ) , return_tensors="""pt""" ).pixel_values # verify logits lowercase = torch.tensor([[model.config.decoder.decoder_start_token_id]] ) lowercase = model(pixel_values=lowerCAmelCase__ , decoder_input_ids=lowerCAmelCase__ ) lowercase = outputs.logits lowercase = torch.Size([1, 1, 5_0_2_6_5] ) if "trocr-base-handwritten" in checkpoint_url: lowercase = torch.tensor( [-1.4_502, -4.6_683, -0.5_347, -2.9_291, 9.1_435, -3.0_571, 8.9_764, 1.7_560, 8.7_358, -1.5_311] ) elif "trocr-large-handwritten" in checkpoint_url: lowercase = torch.tensor( [-2.6_437, -1.3_129, -2.2_596, -5.3_455, 6.3_539, 1.7_604, 5.4_991, 1.4_702, 5.6_113, 2.0_170] ) elif "trocr-base-printed" in checkpoint_url: lowercase = torch.tensor( [-5.6_816, -5.8_388, 1.1_398, -6.9_034, 6.8_505, -2.4_393, 1.2_284, -1.0_232, -1.9_661, -3.9_210] ) elif "trocr-large-printed" in checkpoint_url: lowercase = torch.tensor( [-6.0_162, -7.0_959, 4.4_155, -5.1_063, 7.0_468, -3.1_631, 2.6_466, -0.3_081, -0.8_106, -1.7_535] ) if "stage1" not in checkpoint_url: assert logits.shape == expected_shape, "Shape of logits not as expected" assert torch.allclose(logits[0, 0, :1_0] , lowerCAmelCase__ , atol=1e-3 ), "First elements of logits not as expected" Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(lowerCAmelCase__ ) print(f'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] =argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/trocr/model_zoo/fairseq/trocr-base-handwritten.pt""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) __lowerCAmelCase : Dict =parser.parse_args() convert_tr_ocr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
32
1
"""simple docstring""" import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def lowerCAmelCase__ ( _UpperCamelCase : List[Any] , _UpperCamelCase : Any ) -> Dict: """simple docstring""" snake_case = XCLIPTextConfig() # derive patch size from model name snake_case = model_name.find('patch' ) snake_case = int(model_name[start_idx + len('patch' ) : start_idx + len('patch' ) + 2] ) snake_case = XCLIPVisionConfig(patch_size=_UpperCamelCase , num_frames=_UpperCamelCase ) if "large" in model_name: snake_case = 7_6_8 snake_case = 3_0_7_2 snake_case = 1_2 snake_case = 1_0_2_4 snake_case = 4_0_9_6 snake_case = 1_6 snake_case = 2_4 snake_case = 7_6_8 snake_case = 3_0_7_2 if model_name == "xclip-large-patch14-16-frames": snake_case = 3_3_6 snake_case = XCLIPConfig.from_text_vision_configs(_UpperCamelCase , _UpperCamelCase ) if "large" in model_name: snake_case = 7_6_8 return config def lowerCAmelCase__ ( _UpperCamelCase : Dict ) -> List[Any]: """simple docstring""" if name == "token_embedding.weight": snake_case = name.replace('token_embedding.weight' , 'text_model.embeddings.token_embedding.weight' ) if name == "positional_embedding": snake_case = name.replace('positional_embedding' , 'text_model.embeddings.position_embedding.weight' ) if "ln_1" in name: snake_case = name.replace('ln_1' , 'layer_norm1' ) if "ln_2" in name: snake_case = name.replace('ln_2' , 'layer_norm2' ) if "c_fc" in name: snake_case = name.replace('c_fc' , 'fc1' ) if "c_proj" in name: snake_case = name.replace('c_proj' , 'fc2' ) if name.startswith('transformer.resblocks' ): snake_case = name.replace('transformer.resblocks' , 'text_model.encoder.layers' ) if "attn.out_proj" in name and "message" not in name: snake_case = name.replace('attn.out_proj' , 'self_attn.out_proj' ) if "ln_final" in name: snake_case = name.replace('ln_final' , 'text_model.final_layer_norm' ) # visual encoder if name == "visual.class_embedding": snake_case = name.replace('visual.class_embedding' , 'vision_model.embeddings.class_embedding' ) if name == "visual.positional_embedding": snake_case = name.replace('visual.positional_embedding' , 'vision_model.embeddings.position_embedding.weight' ) if name.startswith('visual.transformer.resblocks' ): snake_case = name.replace('visual.transformer.resblocks' , 'vision_model.encoder.layers' ) if "visual.conv1" in name: snake_case = name.replace('visual.conv1' , 'vision_model.embeddings.patch_embedding' ) if "visual.ln_pre" in name: snake_case = name.replace('visual.ln_pre' , 'vision_model.pre_layernorm' ) if "visual.ln_post" in name: snake_case = name.replace('visual.ln_post' , 'vision_model.post_layernorm' ) if "visual.proj" in name: snake_case = name.replace('visual.proj' , 'visual_projection.weight' ) if "text_projection" in name: snake_case = name.replace('text_projection' , 'text_projection.weight' ) # things on top if "prompts_visual_proj" in name: snake_case = name.replace('prompts_visual_proj' , 'prompts_visual_projection' ) if "prompts_visual_ln" in name: snake_case = name.replace('prompts_visual_ln' , 'prompts_visual_layernorm' ) # mit if name == "mit.positional_embedding": snake_case = name.replace('positional' , 'position' ) if name.startswith('mit.resblocks' ): snake_case = name.replace('mit.resblocks' , 'mit.encoder.layers' ) # prompts generator if name.startswith('prompts_generator.norm' ): snake_case = name.replace('prompts_generator.norm' , 'prompts_generator.layernorm' ) return name def lowerCAmelCase__ ( _UpperCamelCase : List[Any] , _UpperCamelCase : List[str] ) -> List[str]: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case = orig_state_dict.pop(_UpperCamelCase ) if "attn.in_proj" in key: snake_case = key.split('.' ) if key.startswith('visual' ): snake_case = key_split[3] snake_case = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: snake_case = val[ :dim, : ] snake_case = val[ dim : dim * 2, : ] snake_case = val[ -dim:, : ] else: snake_case = val[ :dim ] snake_case = val[ dim : dim * 2 ] snake_case = val[ -dim: ] else: if "weight" in key: snake_case = val[ :dim, : ] snake_case = val[ dim : dim * 2, : ] snake_case = val[ -dim:, : ] else: snake_case = val[:dim] snake_case = val[ dim : dim * 2 ] snake_case = val[-dim:] elif key.startswith('mit' ): snake_case = key_split[2] snake_case = config.vision_config.mit_hidden_size if "weight" in key: snake_case = val[:dim, :] snake_case = val[dim : dim * 2, :] snake_case = val[-dim:, :] else: snake_case = val[:dim] snake_case = val[dim : dim * 2] snake_case = val[-dim:] else: snake_case = key_split[2] snake_case = config.text_config.hidden_size if "weight" in key: snake_case = val[:dim, :] snake_case = val[ dim : dim * 2, : ] snake_case = val[-dim:, :] else: snake_case = val[:dim] snake_case = val[ dim : dim * 2 ] snake_case = val[-dim:] else: snake_case = rename_key(_UpperCamelCase ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: snake_case = val.T snake_case = val return orig_state_dict def lowerCAmelCase__ ( _UpperCamelCase : List[str] ) -> Optional[Any]: """simple docstring""" if num_frames == 8: snake_case = 'eating_spaghetti_8_frames.npy' elif num_frames == 1_6: snake_case = 'eating_spaghetti.npy' elif num_frames == 3_2: snake_case = 'eating_spaghetti_32_frames.npy' snake_case = hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename=_UpperCamelCase , repo_type='dataset' , ) snake_case = np.load(_UpperCamelCase ) return list(_UpperCamelCase ) def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : Tuple=None , _UpperCamelCase : Union[str, Any]=False ) -> Optional[int]: """simple docstring""" snake_case = { # fully supervised kinetics-400 checkpoints 'xclip-base-patch32': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth', 'xclip-base-patch32-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth' ), 'xclip-base-patch16': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth', 'xclip-base-patch16-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth' ), 'xclip-large-patch14': 'https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb', 'xclip-large-patch14-16-frames': 'https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f', # fully supervised kinetics-600 checkpoints 'xclip-base-patch16-kinetics-600': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth' ), 'xclip-base-patch16-kinetics-600-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth' ), 'xclip-large-patch14-kinetics-600': 'https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be', # few shot 'xclip-base-patch16-hmdb-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth' ), 'xclip-base-patch16-hmdb-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth' ), 'xclip-base-patch16-hmdb-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth' ), 'xclip-base-patch16-hmdb-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth' ), 'xclip-base-patch16-ucf-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth' ), 'xclip-base-patch16-ucf-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth' ), 'xclip-base-patch16-ucf-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth' ), 'xclip-base-patch16-ucf-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth' ), # zero shot 'xclip-base-patch16-zero-shot': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth', } snake_case = model_to_url[model_name] snake_case = 8 if "16-frames" in model_name: snake_case = 1_6 elif "shot" in model_name: snake_case = 3_2 snake_case = get_xclip_config(_UpperCamelCase , _UpperCamelCase ) snake_case = XCLIPModel(_UpperCamelCase ) model.eval() if "drive" in checkpoint_url: snake_case = 'pytorch_model.bin' gdown.cached_download(_UpperCamelCase , _UpperCamelCase , quiet=_UpperCamelCase ) snake_case = torch.load(_UpperCamelCase , map_location='cpu' )['model'] else: snake_case = torch.hub.load_state_dict_from_url(_UpperCamelCase )['model'] snake_case = convert_state_dict(_UpperCamelCase , _UpperCamelCase ) snake_case = XCLIPModel(_UpperCamelCase ) snake_case ,snake_case = model.load_state_dict(_UpperCamelCase , strict=_UpperCamelCase ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() snake_case = 3_3_6 if model_name == 'xclip-large-patch14-16-frames' else 2_2_4 snake_case = VideoMAEImageProcessor(size=_UpperCamelCase ) snake_case = CLIPTokenizer.from_pretrained('openai/clip-vit-base-patch32' ) snake_case = CLIPTokenizerFast.from_pretrained('openai/clip-vit-base-patch32' ) snake_case = XCLIPProcessor(image_processor=_UpperCamelCase , tokenizer=_UpperCamelCase ) snake_case = prepare_video(_UpperCamelCase ) snake_case = processor( text=['playing sports', 'eating spaghetti', 'go shopping'] , videos=_UpperCamelCase , return_tensors='pt' , padding=_UpperCamelCase ) print('Shape of pixel values:' , inputs.pixel_values.shape ) with torch.no_grad(): snake_case = model(**_UpperCamelCase ) # Verify outputs snake_case = outputs.logits_per_video snake_case = logits_per_video.softmax(dim=1 ) print('Probs:' , _UpperCamelCase ) # kinetics-400 if model_name == "xclip-base-patch32": snake_case = torch.tensor([[0.00_19, 0.99_51, 0.00_30]] ) elif model_name == "xclip-base-patch32-16-frames": snake_case = torch.tensor([[7.0_9_9_9e-0_4, 9.9_8_8_3e-0_1, 4.5_5_8_0e-0_4]] ) elif model_name == "xclip-base-patch16": snake_case = torch.tensor([[0.00_83, 0.96_81, 0.02_36]] ) elif model_name == "xclip-base-patch16-16-frames": snake_case = torch.tensor([[7.6_9_3_7e-0_4, 9.9_7_2_8e-0_1, 1.9_4_7_3e-0_3]] ) elif model_name == "xclip-large-patch14": snake_case = torch.tensor([[0.00_62, 0.98_64, 0.00_75]] ) elif model_name == "xclip-large-patch14-16-frames": snake_case = torch.tensor([[3.3_8_7_7e-0_4, 9.9_9_3_7e-0_1, 2.8_8_8_8e-0_4]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": snake_case = torch.tensor([[0.05_55, 0.89_14, 0.05_31]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": snake_case = torch.tensor([[3.8_5_5_4e-0_4, 9.9_9_2_9e-0_1, 3.2_7_5_4e-0_4]] ) elif model_name == "xclip-large-patch14-kinetics-600": snake_case = torch.tensor([[0.00_36, 0.99_20, 0.00_45]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": snake_case = torch.tensor([[7.1_8_9_0e-0_6, 9.9_9_9_4e-0_1, 5.6_5_5_9e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": snake_case = torch.tensor([[1.0_3_2_0e-0_5, 9.9_9_9_3e-0_1, 6.2_4_3_5e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": snake_case = torch.tensor([[4.1_3_7_7e-0_6, 9.9_9_9_0e-0_1, 9.8_3_8_6e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": snake_case = torch.tensor([[4.1_3_4_7e-0_5, 9.9_9_6_2e-0_1, 3.3_4_1_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": snake_case = torch.tensor([[8.5_8_5_7e-0_5, 9.9_9_2_8e-0_1, 6.3_2_9_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": snake_case = torch.tensor([[8.5_8_5_7e-0_5, 9.9_9_2_8e-0_1, 6.3_2_9_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": snake_case = torch.tensor([[0.00_27, 0.99_04, 0.00_70]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": snake_case = torch.tensor([[9.8_2_1_9e-0_4, 9.9_5_9_3e-0_1, 3.0_8_6_3e-0_3]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": snake_case = torch.tensor([[3.5_0_8_2e-0_4, 9.9_7_8_5e-0_1, 1.7_9_6_6e-0_3]] ) else: raise ValueError(f"""Model name {model_name} not supported""" ) assert torch.allclose(_UpperCamelCase , _UpperCamelCase , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCamelCase ) if push_to_hub: print('Pushing model, processor and slow tokenizer files to the hub...' ) model.push_to_hub(_UpperCamelCase , organization='nielsr' ) processor.push_to_hub(_UpperCamelCase , organization='nielsr' ) slow_tokenizer.push_to_hub(_UpperCamelCase , organization='nielsr' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="xclip-base-patch32", type=str, help="Name of the model.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
150
"""simple docstring""" import os import jsonlines import numpy as np from tqdm import tqdm SCREAMING_SNAKE_CASE__ = 2_048 SCREAMING_SNAKE_CASE__ = 4_096 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = os.environ.pop("PROCESS_TRAIN", "false") SCREAMING_SNAKE_CASE__ = {"null": 0, "short": 1, "long": 2, "yes": 3, "no": 4} def lowerCAmelCase__ ( _UpperCamelCase : str ) -> Any: """simple docstring""" def choose_first(_UpperCamelCase : Tuple , _UpperCamelCase : List[Any]=False ): assert isinstance(_UpperCamelCase , _UpperCamelCase ) if len(_UpperCamelCase ) == 1: snake_case = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: snake_case = {k: [a[k]] for k in a} if len(a['start_token'] ) > 0: break return a snake_case = {'id': example['id']} snake_case = example['annotations'] snake_case = annotation['yes_no_answer'] if 0 in yes_no_answer or 1 in yes_no_answer: snake_case = ['yes'] if 1 in yes_no_answer else ['no'] snake_case = snake_case = [] snake_case = snake_case = [] snake_case = ['<cls>'] else: snake_case = ['short'] snake_case = choose_first(annotation['short_answers'] ) if len(out['start_token'] ) == 0: # answer will be long if short is not available snake_case = ['long'] snake_case = choose_first(annotation['long_answer'] , is_long_answer=_UpperCamelCase ) snake_case = [] answer.update(_UpperCamelCase ) # disregard some samples if len(answer['start_token'] ) > 1 or answer["start_token"] == answer["end_token"]: snake_case = True else: snake_case = False snake_case = ['start_token', 'end_token', 'start_byte', 'end_byte', 'text'] if not all(isinstance(answer[k] , _UpperCamelCase ) for k in cols ): raise ValueError('Issue in ID' , example['id'] ) return answer def lowerCAmelCase__ ( _UpperCamelCase : Dict , _UpperCamelCase : List[str]=False ) -> Union[str, Any]: """simple docstring""" snake_case = _get_single_answer(_UpperCamelCase ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element snake_case = example['document']['tokens'] snake_case = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) return { "context": " ".join(_UpperCamelCase ), "answer": { "start_token": -1_0_0, # ignore index in cross-entropy "end_token": -1_0_0, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples snake_case = ['start_token', 'end_token'] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 snake_case = example['document']['tokens'] snake_case = answer['start_token'] snake_case = answer['end_token'] snake_case = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 snake_case = ' '.join(context[start_token:end_token] ) # checking above code if assertion: snake_case = doc['is_html'][answer['start_token'] : answer['end_token']] snake_case = doc['token'][answer['start_token'] : answer['end_token']] snake_case = ' '.join([old[i] for i in range(len(_UpperCamelCase ) ) if not is_html[i]] ) if new != old: print('ID:' , example['id'] ) print('New:' , _UpperCamelCase , end='\n' ) print('Old:' , _UpperCamelCase , end='\n\n' ) return { "context": " ".join(_UpperCamelCase ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def lowerCAmelCase__ ( _UpperCamelCase : Dict , _UpperCamelCase : Any , _UpperCamelCase : Optional[int]=2_0_4_8 , _UpperCamelCase : Union[str, Any]=4_0_9_6 , _UpperCamelCase : Dict=True ) -> Optional[Any]: """simple docstring""" snake_case = get_context_and_ans(_UpperCamelCase , assertion=_UpperCamelCase ) snake_case = out['answer'] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } snake_case = tokenizer(example['question']['text'] , out['context'] ).input_ids snake_case = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element snake_case = [] snake_case = [] snake_case = input_ids[:q_len] snake_case = range(_UpperCamelCase , len(_UpperCamelCase ) , max_length - doc_stride ) for i in doc_start_indices: snake_case = i + max_length - q_len snake_case = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['category'][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-1_0_0] * len(_UpperCamelCase ), "end_token": [-1_0_0] * len(_UpperCamelCase ), "category": category, }, } snake_case = out['context'].split() snake_case = splitted_context[answer['end_token']] snake_case = len( tokenizer( ' '.join(splitted_context[: answer['start_token']] ) , add_special_tokens=_UpperCamelCase , ).input_ids ) snake_case = len( tokenizer(' '.join(splitted_context[: answer['end_token']] ) , add_special_tokens=_UpperCamelCase ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token snake_case = len(tokenizer(_UpperCamelCase , add_special_tokens=_UpperCamelCase ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 snake_case = input_ids[answer['start_token'] : answer['end_token'] + 1] # right & left are inclusive snake_case = answer['start_token'] snake_case = answer['end_token'] if assertion: snake_case = tokenizer.decode(_UpperCamelCase ) if answer["span"] != new: print('ISSUE IN TOKENIZATION' ) print('OLD:' , answer['span'] ) print('NEW:' , _UpperCamelCase , end='\n\n' ) if len(_UpperCamelCase ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } snake_case = input_ids[:q_len] snake_case = range(_UpperCamelCase , len(_UpperCamelCase ) , max_length - doc_stride ) snake_case = [] snake_case = [] snake_case = [] snake_case = [] # null, yes, no, long, short for i in doc_start_indices: snake_case = i + max_length - q_len snake_case = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: snake_case = start_token - i + q_len snake_case = end_token - i + q_len answers_category.append(answer['category'][0] ) # ["short"] -> "short" else: snake_case = -1_0_0 snake_case = -1_0_0 answers_category.append('null' ) snake_case = inputs[-1][start_token : end_token + 1] answers_start_token.append(_UpperCamelCase ) answers_end_token.append(_UpperCamelCase ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('ISSUE in strided for ID:' , example['id'] ) print('New:' , tokenizer.decode(_UpperCamelCase ) ) print('Old:' , tokenizer.decode(_UpperCamelCase ) , end='\n\n' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def lowerCAmelCase__ ( _UpperCamelCase : Optional[Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int]=2_0_4_8 , _UpperCamelCase : Union[str, Any]=4_0_9_6 , _UpperCamelCase : List[str]=False ) -> Union[str, Any]: """simple docstring""" snake_case = get_strided_contexts_and_ans( _UpperCamelCase , _UpperCamelCase , doc_stride=_UpperCamelCase , max_length=_UpperCamelCase , assertion=_UpperCamelCase , ) return example def lowerCAmelCase__ ( _UpperCamelCase : Tuple , _UpperCamelCase : List[str] ) -> Any: """simple docstring""" with jsonlines.open(_UpperCamelCase , 'a' ) as writer: for example in tqdm(_UpperCamelCase , total=len(_UpperCamelCase ) , desc='Saving samples ... ' ): snake_case = example['labels'] for ids, start, end, cat in zip( example['input_ids'] , labels['start_token'] , labels['end_token'] , labels['category'] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { 'input_ids': ids, 'start_token': start, 'end_token': end, 'category': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer SCREAMING_SNAKE_CASE__ = load_dataset("natural_questions") SCREAMING_SNAKE_CASE__ = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base") SCREAMING_SNAKE_CASE__ = data["train" if PROCESS_TRAIN == "true" else "validation"] SCREAMING_SNAKE_CASE__ = { "tokenizer": tokenizer, "doc_stride": DOC_STRIDE, "max_length": MAX_LENGTH, "assertion": False, } SCREAMING_SNAKE_CASE__ = data.map(prepare_inputs, fn_kwargs=fn_kwargs) SCREAMING_SNAKE_CASE__ = data.remove_columns(["annotations", "document", "id", "question"]) print(data) np.random.seed(SEED) SCREAMING_SNAKE_CASE__ = "nq-training.jsonl" if PROCESS_TRAIN == "true" else "nq-validation.jsonl" save_to_disk(data, file_name=cache_file_name)
150
1
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class a__ : _a : int _a : TreeNode | None = None _a : TreeNode | None = None UpperCamelCase__ = namedtuple("""CoinsDistribResult""", """moves excess""") def _a ( SCREAMING_SNAKE_CASE_ : TreeNode | None ): if root is None: return 0 # Validation def count_nodes(SCREAMING_SNAKE_CASE_ : TreeNode | None ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(SCREAMING_SNAKE_CASE_ : TreeNode | None ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(SCREAMING_SNAKE_CASE_ ) != count_coins(SCREAMING_SNAKE_CASE_ ): raise ValueError("The nodes number should be same as the number of coins" ) # Main calculation def get_distrib(SCREAMING_SNAKE_CASE_ : TreeNode | None ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) __lowerCAmelCase , __lowerCAmelCase = get_distrib(node.left ) __lowerCAmelCase , __lowerCAmelCase = get_distrib(node.right ) __lowerCAmelCase = 1 - left_distrib_excess __lowerCAmelCase = 1 - right_distrib_excess __lowerCAmelCase = ( left_distrib_moves + right_distrib_moves + abs(SCREAMING_SNAKE_CASE_ ) + abs(SCREAMING_SNAKE_CASE_ ) ) __lowerCAmelCase = node.data - coins_to_left - coins_to_right return CoinsDistribResult(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return get_distrib(SCREAMING_SNAKE_CASE_ )[0] if __name__ == "__main__": import doctest doctest.testmod()
363
from sklearn.metrics import mean_squared_error import datasets UpperCamelCase__ = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ UpperCamelCase__ = """\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. """ UpperCamelCase__ = """ Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {\"raw_values\", \"uniform_average\"} or array-like of shape (n_outputs,), default=\"uniform_average\" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. \"raw_values\" : Returns a full set of errors in case of multioutput input. \"uniform_average\" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric(\"mse\") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {'mse': 0.6123724356957945} If you're using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric(\"mse\", \"multilist\") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput='raw_values') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mse': array([0.41666667, 1. ])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ "https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html" ] , ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("float" ) ), "references": datasets.Sequence(datasets.Value("float" ) ), } else: return { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } def __SCREAMING_SNAKE_CASE( self , _A , _A , _A=None , _A="uniform_average" , _A=True ): """simple docstring""" __lowerCAmelCase = mean_squared_error( _A , _A , sample_weight=_A , multioutput=_A , squared=_A ) return {"mse": mse}
102
0
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'vocab_file': 'spiece.model'} lowerCAmelCase__ = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', } } lowerCAmelCase__ = { 'albert-base-v1': 5_12, 'albert-large-v1': 5_12, 'albert-xlarge-v1': 5_12, 'albert-xxlarge-v1': 5_12, 'albert-base-v2': 5_12, 'albert-large-v2': 5_12, 'albert-xlarge-v2': 5_12, 'albert-xxlarge-v2': 5_12, } lowerCAmelCase__ = '▁' class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __lowerCamelCase , __lowerCamelCase=True , __lowerCamelCase=True , __lowerCamelCase=False , __lowerCamelCase="[CLS]" , __lowerCamelCase="[SEP]" , __lowerCamelCase="<unk>" , __lowerCamelCase="[SEP]" , __lowerCamelCase="<pad>" , __lowerCamelCase="[CLS]" , __lowerCamelCase="[MASK]" , __lowerCamelCase = None , **__lowerCamelCase , ) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _A : Tuple = ( AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase , normalized=__lowerCamelCase) if isinstance(__lowerCamelCase , __lowerCamelCase) else mask_token ) _A : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__lowerCamelCase , remove_space=__lowerCamelCase , keep_accents=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , sep_token=__lowerCamelCase , pad_token=__lowerCamelCase , cls_token=__lowerCamelCase , mask_token=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCamelCase , ) _A : Union[str, Any] = do_lower_case _A : List[str] = remove_space _A : str = keep_accents _A : List[str] = vocab_file _A : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(__lowerCamelCase) @property def _lowerCamelCase ( self) -> Tuple: return len(self.sp_model) def _lowerCamelCase ( self) -> List[Any]: _A : Optional[Any] = {self.convert_ids_to_tokens(__lowerCamelCase): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Union[str, Any]: _A : int = self.__dict__.copy() _A : Optional[int] = None return state def __setstate__( self , __lowerCamelCase) -> int: _A : List[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _A : Tuple = {} _A : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def _lowerCamelCase ( self , __lowerCamelCase) -> int: if self.remove_space: _A : Optional[Any] = " ".join(inputs.strip().split()) else: _A : List[str] = inputs _A : int = outputs.replace("``" , "\"").replace("''" , "\"") if not self.keep_accents: _A : Dict = unicodedata.normalize("NFKD" , __lowerCamelCase) _A : Union[str, Any] = "".join([c for c in outputs if not unicodedata.combining(__lowerCamelCase)]) if self.do_lower_case: _A : Optional[Any] = outputs.lower() return outputs def _lowerCamelCase ( self , __lowerCamelCase) -> List[str]: _A : Union[str, Any] = self.preprocess_text(__lowerCamelCase) _A : int = self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase) _A : Any = [] for piece in pieces: if len(__lowerCamelCase) > 1 and piece[-1] == str(",") and piece[-2].isdigit(): _A : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(__lowerCamelCase , "")) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0]) == 1: _A : List[str] = cur_pieces[1:] else: _A : Tuple = cur_pieces[0][1:] cur_pieces.append(piece[-1]) new_pieces.extend(__lowerCamelCase) else: new_pieces.append(__lowerCamelCase) return new_pieces def _lowerCamelCase ( self , __lowerCamelCase) -> Any: return self.sp_model.PieceToId(__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> Dict: return self.sp_model.IdToPiece(__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> Tuple: _A : int = [] _A : str = "" _A : List[Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__lowerCamelCase) + token _A : Any = True _A : List[str] = [] else: current_sub_tokens.append(__lowerCamelCase) _A : Union[str, Any] = False out_string += self.sp_model.decode(__lowerCamelCase) return out_string.strip() def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> List[int]: _A : List[str] = [self.sep_token_id] _A : List[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None , __lowerCamelCase = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase) if token_ids_a is not None: return [1] + ([0] * len(__lowerCamelCase)) + [1] + ([0] * len(__lowerCamelCase)) + [1] return [1] + ([0] * len(__lowerCamelCase)) + [1] def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> List[int]: _A : List[str] = [self.sep_token_id] _A : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> Tuple[str]: if not os.path.isdir(__lowerCamelCase): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return _A : int = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(__lowerCamelCase) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , __lowerCamelCase) elif not os.path.isfile(self.vocab_file): with open(__lowerCamelCase , "wb") as fi: _A : List[str] = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase) return (out_vocab_file,)
11
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : Tuple = {'''configuration_wavlm''': ['''WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WavLMConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Dict = [ '''WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WavLMForAudioFrameClassification''', '''WavLMForCTC''', '''WavLMForSequenceClassification''', '''WavLMForXVector''', '''WavLMModel''', '''WavLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys a__ : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
313
0
import math class UpperCAmelCase_ : '''simple docstring''' def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[int]: snake_case_ : List[Any] = 0.0 snake_case_ : str = 0.0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: for i in range(len(_SCREAMING_SNAKE_CASE ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def lowerCAmelCase__ ( ): # Training Examples ( m, n ) snake_case_ : Dict = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) snake_case_ : Dict = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training snake_case_ : Optional[Any] = SelfOrganizingMap() snake_case_ : List[Any] = 3 snake_case_ : List[Any] = 0.5 for _ in range(_lowerCAmelCase ): for j in range(len(_lowerCAmelCase ) ): # training sample snake_case_ : Union[str, Any] = training_samples[j] # Compute the winning vector snake_case_ : int = self_organizing_map.get_winner(_lowerCAmelCase , _lowerCAmelCase ) # Update the winning vector snake_case_ : Optional[Any] = self_organizing_map.update(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # classify test sample snake_case_ : Any = [0, 0, 0, 1] snake_case_ : Union[str, Any] = self_organizing_map.get_winner(_lowerCAmelCase , _lowerCAmelCase ) # results print(F'''Clusters that the test sample belongs to : {winner}''' ) print(F'''Weights that have been trained : {weights}''' ) # running the main() function if __name__ == "__main__": main()
353
import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=512 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=4 , ) -> List[str]: snake_case_ : Dict = parent snake_case_ : List[Any] = batch_size snake_case_ : Union[str, Any] = seq_length snake_case_ : Tuple = is_training snake_case_ : List[str] = use_attention_mask snake_case_ : Union[str, Any] = use_token_type_ids snake_case_ : Optional[Any] = use_labels snake_case_ : Tuple = vocab_size snake_case_ : Dict = hidden_size snake_case_ : List[str] = num_hidden_layers snake_case_ : Tuple = num_attention_heads snake_case_ : Dict = intermediate_size snake_case_ : str = hidden_act snake_case_ : List[Any] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : Optional[Any] = max_position_embeddings snake_case_ : Optional[Any] = type_vocab_size snake_case_ : Union[str, Any] = type_sequence_label_size snake_case_ : str = initializer_range snake_case_ : List[Any] = num_choices def _lowerCAmelCase ( self ) -> Union[str, Any]: snake_case_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : str = None if self.use_attention_mask: snake_case_ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ : Any = None if self.use_token_type_ids: snake_case_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ : Tuple = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _lowerCAmelCase ( self ) -> List[Any]: snake_case_ : int = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ : Any = config_and_inputs snake_case_ : Union[str, Any] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _lowerCAmelCase ( self ) -> Optional[int]: snake_case_ : Union[str, Any] = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ : List[Any] = config_and_inputs snake_case_ : str = True snake_case_ : Union[str, Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' A : List[str] = True A : List[str] = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def _lowerCAmelCase ( self ) -> List[str]: snake_case_ : List[str] = FlaxBertModelTester(self ) @slow def _lowerCAmelCase ( self ) -> Dict: # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. snake_case_ : int = FlaxBertModel.from_pretrained("bert-base-cased" ) snake_case_ : Tuple = model(np.ones((1, 1) ) ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
36
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =None __a =None __a =None __a =None class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" def __init__( self : Tuple , __a : Tuple=1 , __a : List[Any]=0 , __a : Any=2 , __a : Any=5_12 , __a : Optional[Any]="cls" , __a : List[Any]=False , __a : Tuple=True , **__a : Tuple , ): super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) _a = project_dim _a = pooler_fn _a = learn_encoder _a = use_attention_mask class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =[r'pooler', r'logit_scale'] __a =[r'position_ids', r'predictions.decoder.bias'] __a ='roberta' __a =RobertaSeriesConfig def __init__( self : Optional[int] , __a : List[Any] ): super().__init__(__a ) _a = XLMRobertaModel(__a ) _a = nn.Linear(config.hidden_size , config.project_dim ) _a = getattr(__a , "has_pre_transformation" , __a ) if self.has_pre_transformation: _a = nn.Linear(config.hidden_size , config.project_dim ) _a = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def UpperCamelCase__ ( self : Any , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[torch.Tensor] = None , __a : Optional[bool] = None , __a : Optional[bool] = None , __a : Optional[bool] = None , ): _a = return_dict if return_dict is not None else self.config.use_return_dict _a = self.base_model( input_ids=__a , attention_mask=__a , token_type_ids=__a , position_ids=__a , head_mask=__a , inputs_embeds=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , output_attentions=__a , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__a , ) if self.has_pre_transformation: _a = outputs["hidden_states"][-2] _a = self.pre_LN(__a ) _a = self.transformation_pre(__a ) return TransformationModelOutput( projection_state=__a , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: _a = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__a , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
63
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ : List[Any] = { "configuration_resnet": ["RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "ResNetConfig", "ResNetOnnxConfig"] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : int = [ "RESNET_PRETRAINED_MODEL_ARCHIVE_LIST", "ResNetForImageClassification", "ResNetModel", "ResNetPreTrainedModel", "ResNetBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[Any] = [ "TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFResNetForImageClassification", "TFResNetModel", "TFResNetPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = [ "FlaxResNetForImageClassification", "FlaxResNetModel", "FlaxResNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys SCREAMING_SNAKE_CASE__ : Any = _LazyModule(__name__, globals()["__file__"], _import_structure)
270
0
UpperCamelCase_ = { "joule": 1.0, "kilojoule": 1_000, "megajoule": 1_000_000, "gigajoule": 1_000_000_000, "wattsecond": 1.0, "watthour": 3_600, "kilowatthour": 3_600_000, "newtonmeter": 1.0, "calorie_nutr": 4_186.8, "kilocalorie_nutr": 4_186_800.00, "electronvolt": 1.6_0217_6634e-19, "britishthermalunit_it": 1_055.05_585, "footpound": 1.35_58_18, } def A ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> float: '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: UpperCAmelCase_ = ( f"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" f"Valid values are: {', '.join(__UpperCAmelCase )}" ) raise ValueError(__UpperCAmelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
344
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class a_ ( nn.Module ): def __init__( self :Optional[Any]) -> Union[str, Any]: super().__init__() UpperCAmelCase_ = nn.Linear(3 , 4) UpperCAmelCase_ = nn.BatchNormad(4) UpperCAmelCase_ = nn.Linear(4 , 5) def __a ( self :Dict , _lowercase :int) -> str: return self.lineara(self.batchnorm(self.lineara(_lowercase))) class a_ ( _snake_case ): def __a ( self :Tuple , _lowercase :Optional[int] , *_lowercase :Union[str, Any] , **_lowercase :Any) -> Optional[Any]: return (args[0] + 1,) + args[1:], kwargs class a_ ( _snake_case ): def __a ( self :Union[str, Any] , _lowercase :Dict , _lowercase :Tuple) -> int: return output + 1 class a_ ( unittest.TestCase ): def __a ( self :str) -> Optional[int]: UpperCAmelCase_ = ModelForTest() UpperCAmelCase_ = ModelHook() add_hook_to_module(_lowercase , _lowercase) self.assertEqual(test_model._hf_hook , _lowercase) self.assertTrue(hasattr(_lowercase , '''_old_forward''')) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , '''forward''') self.assertListEqual(list(inspect.signature(test_model.forward).parameters) , ['''x''']) remove_hook_from_module(_lowercase) self.assertFalse(hasattr(_lowercase , '''_hf_hook''')) self.assertFalse(hasattr(_lowercase , '''_old_forward''')) def __a ( self :Optional[Any]) -> Any: UpperCAmelCase_ = ModelForTest() UpperCAmelCase_ = ModelHook() add_hook_to_module(_lowercase , _lowercase) add_hook_to_module(_lowercase , _lowercase , append=_lowercase) self.assertEqual(isinstance(test_model._hf_hook , _lowercase) , _lowercase) self.assertEqual(len(test_model._hf_hook.hooks) , 2) self.assertTrue(hasattr(_lowercase , '''_old_forward''')) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , '''forward''') self.assertListEqual(list(inspect.signature(test_model.forward).parameters) , ['''x''']) remove_hook_from_module(_lowercase) self.assertFalse(hasattr(_lowercase , '''_hf_hook''')) self.assertFalse(hasattr(_lowercase , '''_old_forward''')) def __a ( self :Optional[int]) -> Optional[int]: UpperCAmelCase_ = ModelForTest() UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = test_model(x + 1) UpperCAmelCase_ = test_model(x + 2) UpperCAmelCase_ = PreForwardHook() add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) self.assertTrue(torch.allclose(_lowercase , _lowercase , atol=1E-5)) # Attaching a hook to a model when it already has one replaces, does not chain UpperCAmelCase_ = PreForwardHook() add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) self.assertTrue(torch.allclose(_lowercase , _lowercase , atol=1E-5)) # You need to use the sequential hook to chain two or more hooks UpperCAmelCase_ = SequentialHook(PreForwardHook() , PreForwardHook()) add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) assert torch.allclose(_lowercase , _lowercase , atol=1E-5) def __a ( self :List[str]) -> int: UpperCAmelCase_ = ModelForTest() UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = test_model(_lowercase) UpperCAmelCase_ = PostForwardHook() add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) self.assertTrue(torch.allclose(_lowercase , output + 1 , atol=1E-5)) # Attaching a hook to a model when it already has one replaces, does not chain UpperCAmelCase_ = PostForwardHook() add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) self.assertTrue(torch.allclose(_lowercase , output + 1 , atol=1E-5)) # You need to use the sequential hook to chain two or more hooks UpperCAmelCase_ = SequentialHook(PostForwardHook() , PostForwardHook()) add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) assert torch.allclose(_lowercase , output + 2 , atol=1E-5) def __a ( self :str) -> List[Any]: UpperCAmelCase_ = ModelForTest() UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = test_model(_lowercase) UpperCAmelCase_ = PostForwardHook() add_hook_to_module(_lowercase , _lowercase) UpperCAmelCase_ = test_model(_lowercase) self.assertTrue(torch.allclose(_lowercase , output + 1)) self.assertTrue(outputa.requires_grad) UpperCAmelCase_ = True UpperCAmelCase_ = test_model(_lowercase) self.assertFalse(outputa.requires_grad) @require_multi_gpu def __a ( self :Tuple) -> Optional[int]: UpperCAmelCase_ = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0)) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0)) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1)) self.assertEqual(model.lineara.weight.device , torch.device(0)) self.assertEqual(model.batchnorm.weight.device , torch.device(0)) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0)) self.assertEqual(model.lineara.weight.device , torch.device(1)) # We can still make a forward pass. The input does not need to be on any particular device UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , torch.device(1)) # We can add a general hook to put back output on same device as input. add_hook_to_module(_lowercase , AlignDevicesHook(io_same_device=_lowercase)) UpperCAmelCase_ = torch.randn(2 , 3).to(0) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , torch.device(0)) def __a ( self :str) -> List[Any]: UpperCAmelCase_ = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # This will move each submodule on different devices UpperCAmelCase_ = {'''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True} add_hook_to_module(model.lineara , AlignDevicesHook(**_lowercase)) add_hook_to_module(model.batchnorm , AlignDevicesHook(**_lowercase)) add_hook_to_module(model.lineara , AlignDevicesHook(**_lowercase)) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) # Buffers are not included in the offload by default, so are on the execution device UpperCAmelCase_ = torch.device(hook_kwargs['''execution_device''']) self.assertEqual(model.batchnorm.running_mean.device , _lowercase) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara) remove_hook_from_module(model.batchnorm) remove_hook_from_module(model.lineara) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # Now test with buffers included in the offload UpperCAmelCase_ = { '''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True, '''offload_buffers''': True, } add_hook_to_module(model.lineara , AlignDevicesHook(**_lowercase)) add_hook_to_module(model.batchnorm , AlignDevicesHook(**_lowercase)) add_hook_to_module(model.lineara , AlignDevicesHook(**_lowercase)) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''')) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara) remove_hook_from_module(model.batchnorm) remove_hook_from_module(model.lineara) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) def __a ( self :List[Any]) -> str: UpperCAmelCase_ = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # This will move each submodule on different devices UpperCAmelCase_ = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook(_lowercase , execution_device=_lowercase , offload=_lowercase) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) # Buffers are not included in the offload by default, so are on the execution device UpperCAmelCase_ = torch.device(_lowercase) self.assertEqual(model.batchnorm.running_mean.device , _lowercase) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_lowercase) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # Now test with buffers included in the offload attach_align_device_hook(_lowercase , execution_device=_lowercase , offload=_lowercase , offload_buffers=_lowercase) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''')) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_lowercase) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) def __a ( self :Optional[Any]) -> int: UpperCAmelCase_ = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # This will move each submodule on different devices UpperCAmelCase_ = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook( _lowercase , execution_device=_lowercase , offload=_lowercase , weights_map=model.state_dict()) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) # Buffers are not included in the offload by default, so are on the execution device UpperCAmelCase_ = torch.device(_lowercase) self.assertEqual(model.batchnorm.running_mean.device , _lowercase) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_lowercase) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) # Now test with buffers included in the offload attach_align_device_hook( _lowercase , execution_device=_lowercase , offload=_lowercase , weights_map=model.state_dict() , offload_buffers=_lowercase , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''')) self.assertEqual(model.lineara.weight.device , torch.device('''meta''')) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''')) UpperCAmelCase_ = torch.randn(2 , 3) UpperCAmelCase_ = model(_lowercase) self.assertEqual(output.device , _lowercase) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_lowercase) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''')) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''')) self.assertEqual(model.lineara.weight.device , torch.device('''cpu'''))
344
1
"""simple docstring""" import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :List[str] ) -> Dict: # A mock response for an HTTP head request to emulate server down UpperCAmelCase = mock.Mock() UpperCAmelCase = 5_00 UpperCAmelCase = {} UpperCAmelCase = HTTPError UpperCAmelCase = {} # Download this model to make sure it's in the cache. UpperCAmelCase = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=lowercase_ ) as mock_head: UpperCAmelCase = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def UpperCAmelCase__ ( self :List[Any] ) -> Dict: # A mock response for an HTTP head request to emulate server down UpperCAmelCase = mock.Mock() UpperCAmelCase = 5_00 UpperCAmelCase = {} UpperCAmelCase = HTTPError UpperCAmelCase = {} # Download this model to make sure it's in the cache. UpperCAmelCase = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=lowercase_ ) as mock_head: UpperCAmelCase = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self :Any ) -> Union[str, Any]: # This test is for deprecated behavior and can be removed in v5 try: UpperCAmelCase = tempfile.mktemp() with open(lowercase_ , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , lowercase_ ) UpperCAmelCase = AlbertTokenizer.from_pretrained(lowercase_ ) finally: os.remove(lowercase_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 10_00 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def UpperCAmelCase__ ( self :Tuple ) -> Optional[int]: # This test is for deprecated behavior and can be removed in v5 UpperCAmelCase = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class A_ ( unittest.TestCase ): """simple docstring""" __UpperCamelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def UpperCAmelCase__ ( cls :Tuple ) -> List[Any]: UpperCAmelCase = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def UpperCAmelCase__ ( cls :List[str] ) -> Optional[int]: try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def UpperCAmelCase__ ( self :Any ) -> str: with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizer(lowercase_ ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ , repo_id='test-tokenizer' , push_to_hub=lowercase_ , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def UpperCAmelCase__ ( self :int ) -> Union[str, Any]: with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizer(lowercase_ ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( lowercase_ , repo_id='valid_org/test-tokenizer-org' , push_to_hub=lowercase_ , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def UpperCAmelCase__ ( self :int ) -> Tuple: CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = CustomTokenizer(lowercase_ ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizerFast.from_pretrained(lowercase_ ) bert_tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = CustomTokenizerFast.from_pretrained(lowercase_ ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) UpperCAmelCase = AutoTokenizer.from_pretrained( f"""{USER}/test-dynamic-tokenizer""" , use_fast=lowercase_ , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :List[str] ) -> Tuple: UpperCAmelCase = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def UpperCAmelCase__ ( self :Tuple ) -> str: UpperCAmelCase = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def UpperCAmelCase__ ( self :Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def UpperCAmelCase__ ( self :Any ) -> int: UpperCAmelCase = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCAmelCase__ ( self :Any ) -> Optional[int]: UpperCAmelCase = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCAmelCase__ ( self :int ) -> Union[str, Any]: UpperCAmelCase = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def UpperCAmelCase__ ( self :Optional[int] ) -> List[Any]: UpperCAmelCase = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def UpperCAmelCase__ ( self :Optional[int] ) -> Optional[Any]: # Even if the offsets are wrong, we necessarily output correct string # parts. UpperCAmelCase = Trie() UpperCAmelCase = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(lowercase_ , ['AB', 'C'] )
78
"""simple docstring""" class A_ : """simple docstring""" def __init__( self :List[Any] , lowercase_ :int ) -> None: UpperCAmelCase = size UpperCAmelCase = [0] * size UpperCAmelCase = [0] * size @staticmethod def UpperCAmelCase__ ( lowercase_ :int ) -> int: return index | (index + 1) @staticmethod def UpperCAmelCase__ ( lowercase_ :int ) -> int: return (index & (index + 1)) - 1 def UpperCAmelCase__ ( self :Any , lowercase_ :int , lowercase_ :int ) -> None: UpperCAmelCase = value while index < self.size: UpperCAmelCase = self.get_prev(lowercase_ ) + 1 if current_left_border == index: UpperCAmelCase = value else: UpperCAmelCase = max(lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase = self.get_next(lowercase_ ) def UpperCAmelCase__ ( self :List[str] , lowercase_ :int , lowercase_ :int ) -> int: right -= 1 # Because of right is exclusive UpperCAmelCase = 0 while left <= right: UpperCAmelCase = self.get_prev(lowercase_ ) if left <= current_left: UpperCAmelCase = max(lowercase_ , self.tree[right] ) UpperCAmelCase = current_left else: UpperCAmelCase = max(lowercase_ , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
78
1
'''simple docstring''' import mpmath # for roots of unity import numpy as np class UpperCAmelCase_ : """simple docstring""" def __init__( self : Optional[Any] , snake_case_ : Optional[int]=None , snake_case_ : List[Any]=None ): # Input as list snake_case__ : Any = list(poly_a or [0] )[:] snake_case__ : Tuple = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() snake_case__ : Tuple = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() snake_case__ : int = len(self.polyB ) # Add 0 to make lengths equal a power of 2 snake_case__ : Optional[int] = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform snake_case__ : Union[str, Any] = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product snake_case__ : Union[str, Any] = self.__multiply() def lowerCamelCase ( self : str , snake_case_ : Union[str, Any] ): snake_case__ : Dict = [[x] for x in self.polyA] if which == """A""" else [[x] for x in self.polyB] # Corner case if len(a__ ) <= 1: return dft[0] # snake_case__ : Dict = self.c_max_length // 2 while next_ncol > 0: snake_case__ : str = [[] for i in range(a__ )] snake_case__ : Dict = self.root**next_ncol # First half of next step snake_case__ : int = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a__ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step snake_case__ : List[str] = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a__ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update snake_case__ : str = new_dft snake_case__ : Dict = next_ncol // 2 return dft[0] def lowerCamelCase ( self : int ): snake_case__ : Optional[int] = self.__dft("""A""" ) snake_case__ : Any = self.__dft("""B""" ) snake_case__ : List[str] = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT snake_case__ : Dict = 2 while next_ncol <= self.c_max_length: snake_case__ : Union[str, Any] = [[] for i in range(a__ )] snake_case__ : int = self.root ** (next_ncol // 2) snake_case__ : List[Any] = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update snake_case__ : Tuple = new_inverse_c next_ncol *= 2 # Unpack snake_case__ : int = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1J for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : int ): snake_case__ : Dict = """A = """ + """ + """.join( f"{coef}*x^{i}" for coef, i in enumerate(self.polyA[: self.len_A] ) ) snake_case__ : List[str] = """B = """ + """ + """.join( f"{coef}*x^{i}" for coef, i in enumerate(self.polyB[: self.len_B] ) ) snake_case__ : Union[str, Any] = """A*B = """ + """ + """.join( f"{coef}*x^{i}" for coef, i in enumerate(self.product ) ) return f"{a}\n{b}\n{c}" # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
363
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCAmelCase_ ( _a , _a , unittest.TestCase ): """simple docstring""" lowercase = CycleDiffusionPipeline lowercase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "negative_prompt", "height", "width", "negative_prompt_embeds", } lowercase = PipelineTesterMixin.required_optional_params - {"latents"} lowercase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"source_prompt"} ) lowercase = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self : Dict ): torch.manual_seed(0 ) snake_case__ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) snake_case__ : Any = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , num_train_timesteps=1_000 , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) snake_case__ : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case__ : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) snake_case__ : Union[str, Any] = CLIPTextModel(snake_case_ ) snake_case__ : Dict = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case__ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCamelCase ( self : Optional[int] , snake_case_ : Any , snake_case_ : str=0 ): snake_case__ : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) snake_case__ : List[str] = image / 2 + 0.5 if str(snake_case_ ).startswith("""mps""" ): snake_case__ : Tuple = torch.manual_seed(snake_case_ ) else: snake_case__ : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) snake_case__ : str = { """prompt""": """An astronaut riding an elephant""", """source_prompt""": """An astronaut riding a horse""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """eta""": 0.1, """strength""": 0.8, """guidance_scale""": 3, """source_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def lowerCamelCase ( self : Tuple ): snake_case__ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case__ : Optional[Any] = self.get_dummy_components() snake_case__ : List[Any] = CycleDiffusionPipeline(**snake_case_ ) snake_case__ : List[str] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Tuple = self.get_dummy_inputs(snake_case_ ) snake_case__ : Any = pipe(**snake_case_ ) snake_case__ : Tuple = output.images snake_case__ : Optional[int] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) snake_case__ : List[Any] = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def lowerCamelCase ( self : List[str] ): snake_case__ : List[str] = self.get_dummy_components() for name, module in components.items(): if hasattr(snake_case_ , """half""" ): snake_case__ : Optional[int] = module.half() snake_case__ : List[Any] = CycleDiffusionPipeline(**snake_case_ ) snake_case__ : Union[str, Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : int = self.get_dummy_inputs(snake_case_ ) snake_case__ : List[Any] = pipe(**snake_case_ ) snake_case__ : Any = output.images snake_case__ : Optional[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) snake_case__ : List[Any] = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def lowerCamelCase ( self : List[Any] ): return super().test_save_load_local() @unittest.skip("""non-deterministic pipeline""" ) def lowerCamelCase ( self : int ): return super().test_inference_batch_single_identical() @skip_mps def lowerCamelCase ( self : Dict ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowerCamelCase ( self : int ): return super().test_save_load_optional_components() @skip_mps def lowerCamelCase ( self : Optional[int] ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" def lowerCamelCase ( self : Dict ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : str ): snake_case__ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) snake_case__ : Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy""" ) snake_case__ : Tuple = init_image.resize((512, 512) ) snake_case__ : List[Any] = """CompVis/stable-diffusion-v1-4""" snake_case__ : Tuple = DDIMScheduler.from_pretrained(snake_case_ , subfolder="""scheduler""" ) snake_case__ : Optional[int] = CycleDiffusionPipeline.from_pretrained( snake_case_ , scheduler=snake_case_ , safety_checker=snake_case_ , torch_dtype=torch.floataa , revision="""fp16""" ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() snake_case__ : int = """A black colored car""" snake_case__ : int = """A blue colored car""" snake_case__ : Dict = torch.manual_seed(0 ) snake_case__ : Dict = pipe( prompt=snake_case_ , source_prompt=snake_case_ , image=snake_case_ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=snake_case_ , output_type="""np""" , ) snake_case__ : Union[str, Any] = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def lowerCamelCase ( self : int ): snake_case__ : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) snake_case__ : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy""" ) snake_case__ : Dict = init_image.resize((512, 512) ) snake_case__ : Tuple = """CompVis/stable-diffusion-v1-4""" snake_case__ : List[Any] = DDIMScheduler.from_pretrained(snake_case_ , subfolder="""scheduler""" ) snake_case__ : str = CycleDiffusionPipeline.from_pretrained(snake_case_ , scheduler=snake_case_ , safety_checker=snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() snake_case__ : Tuple = """A black colored car""" snake_case__ : List[Any] = """A blue colored car""" snake_case__ : Optional[Any] = torch.manual_seed(0 ) snake_case__ : Any = pipe( prompt=snake_case_ , source_prompt=snake_case_ , image=snake_case_ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=snake_case_ , output_type="""np""" , ) snake_case__ : List[Any] = output.images assert np.abs(image - expected_image ).max() < 2E-2
43
0
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ ( lowerCamelCase__ ): '''simple docstring''' lowerCamelCase__ : int = (PNDMScheduler,) lowerCamelCase__ : Any = (('num_inference_steps', 50),) def _UpperCAmelCase ( self, **lowercase_ ) -> Any: """simple docstring""" a__ ={ '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**lowercase_ ) return config def _UpperCAmelCase ( self, lowercase_=0, **lowercase_ ) -> Tuple: """simple docstring""" a__ =dict(self.forward_default_kwargs ) a__ =kwargs.pop('''num_inference_steps''', lowercase_ ) a__ =self.dummy_sample a__ =0.1 * sample a__ =[residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: a__ =self.get_scheduler_config(**lowercase_ ) a__ =scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals a__ =dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) a__ =scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals a__ =dummy_past_residuals[:] a__ =scheduler.step_prk(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample a__ =new_scheduler.step_prk(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" a__ =scheduler.step_plms(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample a__ =new_scheduler.step_plms(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _UpperCAmelCase ( self ) -> int: """simple docstring""" pass def _UpperCAmelCase ( self, lowercase_=0, **lowercase_ ) -> Any: """simple docstring""" a__ =dict(self.forward_default_kwargs ) a__ =kwargs.pop('''num_inference_steps''', lowercase_ ) a__ =self.dummy_sample a__ =0.1 * sample a__ =[residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: a__ =self.get_scheduler_config() a__ =scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) a__ =dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) a__ =scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) a__ =dummy_past_residuals[:] a__ =scheduler.step_prk(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample a__ =new_scheduler.step_prk(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" a__ =scheduler.step_plms(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample a__ =new_scheduler.step_plms(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _UpperCAmelCase ( self, **lowercase_ ) -> Dict: """simple docstring""" a__ =self.scheduler_classes[0] a__ =self.get_scheduler_config(**lowercase_ ) a__ =scheduler_class(**lowercase_ ) a__ =10 a__ =self.dummy_model() a__ =self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): a__ =model(lowercase_, lowercase_ ) a__ =scheduler.step_prk(lowercase_, lowercase_, lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): a__ =model(lowercase_, lowercase_ ) a__ =scheduler.step_plms(lowercase_, lowercase_, lowercase_ ).prev_sample return sample def _UpperCAmelCase ( self ) -> Optional[int]: """simple docstring""" a__ =dict(self.forward_default_kwargs ) a__ =kwargs.pop('''num_inference_steps''', lowercase_ ) for scheduler_class in self.scheduler_classes: a__ =self.get_scheduler_config() a__ =scheduler_class(**lowercase_ ) a__ =self.dummy_sample a__ =0.1 * sample if num_inference_steps is not None and hasattr(lowercase_, '''set_timesteps''' ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_, '''set_timesteps''' ): a__ =num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) a__ =[residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] a__ =dummy_past_residuals[:] a__ =scheduler.step_prk(lowercase_, 0, lowercase_, **lowercase_ ).prev_sample a__ =scheduler.step_prk(lowercase_, 1, lowercase_, **lowercase_ ).prev_sample self.assertEqual(output_a.shape, sample.shape ) self.assertEqual(output_a.shape, output_a.shape ) a__ =scheduler.step_plms(lowercase_, 0, lowercase_, **lowercase_ ).prev_sample a__ =scheduler.step_plms(lowercase_, 1, lowercase_, **lowercase_ ).prev_sample self.assertEqual(output_a.shape, sample.shape ) self.assertEqual(output_a.shape, output_a.shape ) def _UpperCAmelCase ( self ) -> str: """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def _UpperCAmelCase ( self ) -> Optional[int]: """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) a__ =self.scheduler_classes[0] a__ =self.get_scheduler_config(steps_offset=1 ) a__ =scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps, torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ), ) def _UpperCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" for beta_start, beta_end in zip([0.0001, 0.001], [0.002, 0.02] ): self.check_over_configs(beta_start=lowercase_, beta_end=lowercase_ ) def _UpperCAmelCase ( self ) -> Optional[Any]: """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def _UpperCAmelCase ( self ) -> List[str]: """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def _UpperCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def _UpperCAmelCase ( self ) -> str: """simple docstring""" for t, num_inference_steps in zip([1, 5, 10], [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def _UpperCAmelCase ( self ) -> List[Any]: """simple docstring""" a__ =27 for scheduler_class in self.scheduler_classes: a__ =self.dummy_sample a__ =0.1 * sample a__ =self.get_scheduler_config() a__ =scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): a__ =scheduler.step_prk(lowercase_, lowercase_, lowercase_ ).prev_sample def _UpperCAmelCase ( self ) -> Dict: """simple docstring""" with self.assertRaises(lowercase_ ): a__ =self.scheduler_classes[0] a__ =self.get_scheduler_config() a__ =scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample, 1, self.dummy_sample ).prev_sample def _UpperCAmelCase ( self ) -> Optional[Any]: """simple docstring""" a__ =self.full_loop() a__ =torch.sum(torch.abs(lowercase_ ) ) a__ =torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 198.1318 ) < 1E-2 assert abs(result_mean.item() - 0.2580 ) < 1E-3 def _UpperCAmelCase ( self ) -> Tuple: """simple docstring""" a__ =self.full_loop(prediction_type='''v_prediction''' ) a__ =torch.sum(torch.abs(lowercase_ ) ) a__ =torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.3986 ) < 1E-2 assert abs(result_mean.item() - 0.0878 ) < 1E-3 def _UpperCAmelCase ( self ) -> Any: """simple docstring""" a__ =self.full_loop(set_alpha_to_one=lowercase_, beta_start=0.01 ) a__ =torch.sum(torch.abs(lowercase_ ) ) a__ =torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 230.0399 ) < 1E-2 assert abs(result_mean.item() - 0.2995 ) < 1E-3 def _UpperCAmelCase ( self ) -> Tuple: """simple docstring""" a__ =self.full_loop(set_alpha_to_one=lowercase_, beta_start=0.01 ) a__ =torch.sum(torch.abs(lowercase_ ) ) a__ =torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 186.9482 ) < 1E-2 assert abs(result_mean.item() - 0.2434 ) < 1E-3
188
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase = { '''configuration_canine''': ['''CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CanineConfig'''], '''tokenization_canine''': ['''CanineTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''CANINE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CanineForMultipleChoice''', '''CanineForQuestionAnswering''', '''CanineForSequenceClassification''', '''CanineForTokenClassification''', '''CanineLayer''', '''CanineModel''', '''CaninePreTrainedModel''', '''load_tf_weights_in_canine''', ] if TYPE_CHECKING: from .configuration_canine import CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP, CanineConfig from .tokenization_canine import CanineTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_canine import ( CANINE_PRETRAINED_MODEL_ARCHIVE_LIST, CanineForMultipleChoice, CanineForQuestionAnswering, CanineForSequenceClassification, CanineForTokenClassification, CanineLayer, CanineModel, CaninePreTrainedModel, load_tf_weights_in_canine, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
188
1
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets lowerCAmelCase = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' lowerCAmelCase = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' lowerCAmelCase = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _a ( datasets.Metric ): def lowerCamelCase_ ( self: Union[str, Any] ) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCamelCase_ ( self: int , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Tuple=None , UpperCamelCase_: Dict=None , UpperCamelCase_: int=None , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: Dict="auto" , UpperCamelCase_: Optional[int]=-1 , UpperCamelCase_: Union[str, Any]=0.9 , UpperCamelCase_: Any=5 , UpperCamelCase_: Optional[Any]=500 , UpperCamelCase_: Any="gpt2-large" , UpperCamelCase_: List[Any]=-1 , UpperCamelCase_: Dict=1_024 , UpperCamelCase_: str=25 , UpperCamelCase_: int=5 , UpperCamelCase_: int=True , UpperCamelCase_: Any=25 , ) -> List[Any]: """simple docstring""" lowercase__ = compute_mauve( p_text=UpperCamelCase_ , q_text=UpperCamelCase_ , p_features=UpperCamelCase_ , q_features=UpperCamelCase_ , p_tokens=UpperCamelCase_ , q_tokens=UpperCamelCase_ , num_buckets=UpperCamelCase_ , pca_max_data=UpperCamelCase_ , kmeans_explained_var=UpperCamelCase_ , kmeans_num_redo=UpperCamelCase_ , kmeans_max_iter=UpperCamelCase_ , featurize_model_name=UpperCamelCase_ , device_id=UpperCamelCase_ , max_text_length=UpperCamelCase_ , divergence_curve_discretization_size=UpperCamelCase_ , mauve_scaling_factor=UpperCamelCase_ , verbose=UpperCamelCase_ , seed=UpperCamelCase_ , ) return out
93
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCAmelCase = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCAmelCase = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCAmelCase = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCAmelCase = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _a ( datasets.Metric ): def lowerCamelCase_ ( self: Optional[Any] ) -> Union[str, Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase_: List[str] ) -> Union[str, Any]: """simple docstring""" import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def lowerCamelCase_ ( self: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Dict=0.9 , UpperCamelCase_: Union[str, Any]=3 , UpperCamelCase_: Optional[int]=0.5 ) -> Dict: """simple docstring""" if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase__ = [ meteor_score.single_meteor_score( word_tokenize(UpperCamelCase_ ) , word_tokenize(UpperCamelCase_ ) , alpha=UpperCamelCase_ , beta=UpperCamelCase_ , gamma=UpperCamelCase_ ) for ref, pred in zip(UpperCamelCase_ , UpperCamelCase_ ) ] else: lowercase__ = [ meteor_score.single_meteor_score(UpperCamelCase_ , UpperCamelCase_ , alpha=UpperCamelCase_ , beta=UpperCamelCase_ , gamma=UpperCamelCase_ ) for ref, pred in zip(UpperCamelCase_ , UpperCamelCase_ ) ] return {"meteor": np.mean(UpperCamelCase_ )}
93
1
from collections.abc import Callable import numpy as np def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): __a = int(np.ceil((x_end - xa) / step_size ) ) __a = np.zeros((n + 1,) ) __a = ya __a = xa for k in range(snake_case__ ): __a = y[k] + step_size * ode_func(snake_case__ , y[k] ) __a = y[k] + ( (step_size / 2) * (ode_func(snake_case__ , y[k] ) + ode_func(x + step_size , snake_case__ )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
49
import math import sys def snake_case ( snake_case__ :int) -> int: if number != int(snake_case__): raise ValueError("""the value of input must be a natural number""") if number < 0: raise ValueError("""the value of input must not be a negative number""") if number == 0: return 1 _A = [-1] * (number + 1) _A = 0 for i in range(1 , number + 1): _A = sys.maxsize _A = int(math.sqrt(snake_case__)) for j in range(1 , root + 1): _A = 1 + answers[i - (j**2)] _A = min(snake_case__ , snake_case__) _A = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
180
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False')) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env') @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 650, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'pytorch', 'script': 'run_ddp.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 600, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'tensorflow', 'script': 'run_tf_dist.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 600, 'eval_accuracy': 0.6, 'eval_loss': 0.7}, }, ]) class snake_case ( unittest.TestCase): def a_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=a__ , ) assert hasattr(self , "env" ) def a_ ( self : List[str] , a__ : List[Any] ) -> Any: '''simple docstring''' _A = F"""{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}""" # distributed data settings _A = {"smdistributed": {"dataparallel": {"enabled": True}}} if self.script != "run_ddp.py" else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=a__ , instance_count=a__ , instance_type=self.instance_type , debugger_hook_config=a__ , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=a__ , py_version="py36" , ) def a_ ( self : List[str] , a__ : Union[str, Any] ) -> Dict: '''simple docstring''' TrainingJobAnalytics(a__ ).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(2,)] ) def a_ ( self : List[Any] , a__ : Optional[Any] ) -> Any: '''simple docstring''' _A = self.create_estimator(a__ ) # run training estimator.fit() # result dataframe _A = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _A = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _A = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _A = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 99_99_99 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , a__ )
352
"""simple docstring""" def a__ ( __lowercase , __lowercase ) -> float: _validate_point(__lowercase ) _validate_point(__lowercase ) if len(__lowercase ) != len(__lowercase ): raise ValueError("Both points must be in the same n-dimensional space" ) return float(sum(abs(a - b ) for a, b in zip(__lowercase , __lowercase ) ) ) def a__ ( __lowercase ) -> None: if point: if isinstance(__lowercase , __lowercase ): for item in point: if not isinstance(__lowercase , (int, float) ): _A = ( "Expected a list of numbers as input, found " f"""{type(__lowercase ).__name__}""" ) raise TypeError(__lowercase ) else: _A = f"""Expected a list of numbers as input, found {type(__lowercase ).__name__}""" raise TypeError(__lowercase ) else: raise ValueError("Missing an input" ) def a__ ( __lowercase , __lowercase ) -> float: _validate_point(__lowercase ) _validate_point(__lowercase ) if len(__lowercase ) != len(__lowercase ): raise ValueError("Both points must be in the same n-dimensional space" ) return float(sum(abs(x - y ) for x, y in zip(__lowercase , __lowercase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
163
0
def SCREAMING_SNAKE_CASE_ ( __A : str , __A : str ) -> str: """simple docstring""" a_ : int = len(__A ) a_ : int = len(__A ) a_ : int = ( first_str_length if first_str_length > second_str_length else second_str_length ) a_ : list = [] for char_count in range(__A ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(__A ) if __name__ == "__main__": print(alternative_string_arrange('AB', 'XYZ'), end=' ')
32
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ : Optional[int] = { 'facebook/mask2former-swin-small-coco-instance': ( 'https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json' ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } UpperCAmelCase_ : List[str] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Any = '''mask2former''' snake_case__ : Any = ['''swin'''] snake_case__ : str = {'''hidden_size''': '''hidden_dim'''} def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Optional[Dict] = None , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 1_0_2_4 , SCREAMING_SNAKE_CASE__ : str = "relu" , SCREAMING_SNAKE_CASE__ : int = 6 , SCREAMING_SNAKE_CASE__ : int = 1_0 , SCREAMING_SNAKE_CASE__ : int = 8 , SCREAMING_SNAKE_CASE__ : float = 0.0 , SCREAMING_SNAKE_CASE__ : int = 2_0_4_8 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : int = 4 , SCREAMING_SNAKE_CASE__ : int = 2_5_5 , SCREAMING_SNAKE_CASE__ : int = 1_0_0 , SCREAMING_SNAKE_CASE__ : float = 0.1 , SCREAMING_SNAKE_CASE__ : float = 2.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : int = 1_2_5_4_4 , SCREAMING_SNAKE_CASE__ : float = 3.0 , SCREAMING_SNAKE_CASE__ : float = 0.75 , SCREAMING_SNAKE_CASE__ : float = 0.02 , SCREAMING_SNAKE_CASE__ : float = 1.0 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : List[int] = [4, 8, 1_6, 3_2] , SCREAMING_SNAKE_CASE__ : bool = None , **SCREAMING_SNAKE_CASE__ : int , ) -> List[Any]: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) a_ : Dict = CONFIG_MAPPING['swin']( image_size=2_2_4 , in_channels=3 , patch_size=4 , embed_dim=9_6 , depths=[2, 2, 1_8, 2] , num_heads=[3, 6, 1_2, 2_4] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=SCREAMING_SNAKE_CASE__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): a_ : Any = backbone_config.pop('model_type' ) a_ : Optional[Any] = CONFIG_MAPPING[backbone_model_type] a_ : List[str] = config_class.from_dict(SCREAMING_SNAKE_CASE__ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. """ F"""Supported model types: {",".join(self.backbones_supported )}""" ) a_ : Dict = backbone_config a_ : List[str] = feature_size a_ : List[str] = mask_feature_size a_ : int = hidden_dim a_ : Dict = encoder_feedforward_dim a_ : str = activation_function a_ : List[str] = encoder_layers a_ : List[str] = decoder_layers a_ : Dict = num_attention_heads a_ : str = dropout a_ : Tuple = dim_feedforward a_ : List[str] = pre_norm a_ : Optional[int] = enforce_input_projection a_ : Any = common_stride a_ : Optional[int] = ignore_value a_ : int = num_queries a_ : Tuple = no_object_weight a_ : Dict = class_weight a_ : Optional[int] = mask_weight a_ : Optional[int] = dice_weight a_ : str = train_num_points a_ : List[str] = oversample_ratio a_ : List[Any] = importance_sample_ratio a_ : Any = init_std a_ : Union[str, Any] = init_xavier_std a_ : Union[str, Any] = use_auxiliary_loss a_ : Dict = feature_strides a_ : List[str] = output_auxiliary_logits a_ : Dict = decoder_layers super().__init__(**SCREAMING_SNAKE_CASE__ ) @classmethod def SCREAMING_SNAKE_CASE ( cls : str , SCREAMING_SNAKE_CASE__ : PretrainedConfig , **SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> List[Any]: return cls( backbone_config=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict[str, any]: a_ : Optional[int] = copy.deepcopy(self.__dict__ ) a_ : List[Any] = self.backbone_config.to_dict() a_ : Optional[Any] = self.__class__.model_type return output
32
1
def __lowerCAmelCase ( a__ ) -> Optional[int]: __a = set() # edges = list of graph's edges __a = get_edges(A__ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: __a = edges.pop() chosen_vertices.add(A__ ) chosen_vertices.add(A__ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(A__ ) return chosen_vertices def __lowerCAmelCase ( a__ ) -> str: __a = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
350
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A : Optional[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Tuple = ['NllbTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[Any] = ['NllbTokenizerFast'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys A : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
33
0
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" lowerCamelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__A , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(__A , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(__A , "num_attention_heads" ) ) class UpperCAmelCase_ : '''simple docstring''' def __init__( self , __A , __A=13 , __A=32 , __A=2 , __A=3 , __A=640 , __A=4 , __A="silu" , __A=3 , __A=32 , __A=0.1 , __A=0.1 , __A=0.1 , __A=0.02 , __A=True , __A=True , __A=10 , __A=None , ): """simple docstring""" lowerCamelCase : str = parent lowerCamelCase : str = batch_size lowerCamelCase : Dict = image_size lowerCamelCase : Dict = patch_size lowerCamelCase : Optional[int] = num_channels lowerCamelCase : int = last_hidden_size lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : int = hidden_act lowerCamelCase : Optional[Any] = conv_kernel_size lowerCamelCase : Tuple = output_stride lowerCamelCase : Any = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Any = classifier_dropout_prob lowerCamelCase : str = use_labels lowerCamelCase : Optional[Any] = is_training lowerCamelCase : str = num_labels lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : Dict = scope def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase : List[str] = None lowerCamelCase : Any = None if self.use_labels: lowerCamelCase : int = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase : List[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def _snake_case ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _snake_case ( self , __A , __A , __A , __A ): """simple docstring""" lowerCamelCase : Optional[Any] = MobileViTModel(config=__A ) model.to(__A ) model.eval() lowerCamelCase : List[str] = model(__A ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _snake_case ( self , __A , __A , __A , __A ): """simple docstring""" lowerCamelCase : str = self.num_labels lowerCamelCase : Union[str, Any] = MobileViTForImageClassification(__A ) model.to(__A ) model.eval() lowerCamelCase : Any = model(__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self , __A , __A , __A , __A ): """simple docstring""" lowerCamelCase : Dict = self.num_labels lowerCamelCase : Any = MobileViTForSemanticSegmentation(__A ) model.to(__A ) model.eval() lowerCamelCase : Optional[Any] = model(__A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowerCamelCase : str = model(__A , labels=__A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[Any] = self.prepare_config_and_inputs() lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : str = config_and_inputs lowerCamelCase : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' __A : int = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A : str = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A : Union[str, Any] = False __A : Dict = False __A : List[Any] = False __A : Union[str, Any] = False def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = MobileViTModelTester(self ) lowerCamelCase : int = MobileViTConfigTester(self , config_class=__A , has_text_modality=__A ) def _snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip(reason="MobileViT does not output attentions" ) def _snake_case ( self ): """simple docstring""" pass def _snake_case ( self ): """simple docstring""" lowerCamelCase , lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : List[str] = model_class(__A ) lowerCamelCase : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase : int = [*signature.parameters.keys()] lowerCamelCase : List[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , __A ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _snake_case ( self ): """simple docstring""" pass def _snake_case ( self ): """simple docstring""" lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def _snake_case ( self ): """simple docstring""" def check_hidden_states_output(__A , __A , __A ): lowerCamelCase : Tuple = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): lowerCamelCase : List[Any] = model(**self._prepare_for_class(__A , __A ) ) lowerCamelCase : Union[str, Any] = outputs.hidden_states lowerCamelCase : Dict = 5 self.assertEqual(len(__A ) , __A ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. lowerCamelCase : Union[str, Any] = 2 for i in range(len(__A ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) lowerCamelCase , lowerCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : Dict = True check_hidden_states_output(__A , __A , __A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase : Any = True check_hidden_states_output(__A , __A , __A ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__A ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__A ) @slow def _snake_case ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase : Dict = MobileViTModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def lowercase_( ): '''simple docstring''' lowerCamelCase : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def _snake_case ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[Any] = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(__A ) lowerCamelCase : Dict = self.default_image_processor lowerCamelCase : List[Any] = prepare_img() lowerCamelCase : Optional[Any] = image_processor(images=__A , return_tensors="pt" ).to(__A ) # forward pass with torch.no_grad(): lowerCamelCase : Union[str, Any] = model(**__A ) # verify the logits lowerCamelCase : Optional[int] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __A ) lowerCamelCase : int = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __A , atol=1e-4 ) ) @slow def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCamelCase : Tuple = model.to(__A ) lowerCamelCase : Tuple = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCamelCase : List[str] = prepare_img() lowerCamelCase : List[Any] = image_processor(images=__A , return_tensors="pt" ).to(__A ) # forward pass with torch.no_grad(): lowerCamelCase : Union[str, Any] = model(**__A ) lowerCamelCase : Any = outputs.logits # verify the logits lowerCamelCase : List[str] = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __A ) lowerCamelCase : Any = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__A , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __A , atol=1e-4 ) ) @slow def _snake_case ( self ): """simple docstring""" lowerCamelCase : Union[str, Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCamelCase : int = model.to(__A ) lowerCamelCase : Optional[Any] = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCamelCase : List[str] = prepare_img() lowerCamelCase : List[str] = image_processor(images=__A , return_tensors="pt" ).to(__A ) # forward pass with torch.no_grad(): lowerCamelCase : Tuple = model(**__A ) lowerCamelCase : int = outputs.logits.detach().cpu() lowerCamelCase : int = image_processor.post_process_semantic_segmentation(outputs=__A , target_sizes=[(50, 60)] ) lowerCamelCase : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __A ) lowerCamelCase : str = image_processor.post_process_semantic_segmentation(outputs=__A ) lowerCamelCase : int = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __A )
283
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' def __init__( self , *__A , **__A ): """simple docstring""" warnings.warn( "The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DPTImageProcessor instead." , __A , ) super().__init__(*__A , **__A )
283
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase : List[str] = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = {'tokenizer_file': 'tokenizer.json'} UpperCAmelCase : str = { 'tokenizer_file': { 'bigscience/tokenizer': 'https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json', 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json', }, } class lowerCamelCase__ ( A ): """simple docstring""" __a = VOCAB_FILES_NAMES __a = PRETRAINED_VOCAB_FILES_MAP __a = ["""input_ids""", """attention_mask"""] __a = None def __init__( self : Optional[int] , UpperCamelCase : Dict=None , UpperCamelCase : Dict=None , UpperCamelCase : Any=None , UpperCamelCase : int="<unk>" , UpperCamelCase : int="<s>" , UpperCamelCase : Union[str, Any]="</s>" , UpperCamelCase : str="<pad>" , UpperCamelCase : Tuple=False , UpperCamelCase : Tuple=False , **UpperCamelCase : Optional[Any] , ): '''simple docstring''' super().__init__( UpperCamelCase , UpperCamelCase , tokenizer_file=UpperCamelCase , unk_token=UpperCamelCase , bos_token=UpperCamelCase , eos_token=UpperCamelCase , pad_token=UpperCamelCase , add_prefix_space=UpperCamelCase , clean_up_tokenization_spaces=UpperCamelCase , **UpperCamelCase , ) __UpperCAmelCase : List[str] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCamelCase ) != add_prefix_space: __UpperCAmelCase : int = getattr(UpperCamelCase , pre_tok_state.pop("""type""" ) ) __UpperCAmelCase : Union[str, Any] = add_prefix_space __UpperCAmelCase : List[Any] = pre_tok_class(**UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = add_prefix_space def lowerCamelCase__ ( self : Optional[Any] , *UpperCamelCase : Optional[int] , **UpperCamelCase : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : str = kwargs.get("""is_split_into_words""" , UpperCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with''' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : int , *UpperCamelCase : Tuple , **UpperCamelCase : List[str] ): '''simple docstring''' __UpperCAmelCase : List[Any] = kwargs.get("""is_split_into_words""" , UpperCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with''' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : List[Any] , UpperCamelCase : str , UpperCamelCase : Optional[str] = None ): '''simple docstring''' __UpperCAmelCase : int = self._tokenizer.model.save(UpperCamelCase , name=UpperCamelCase ) return tuple(UpperCamelCase ) def lowerCamelCase__ ( self : Any , UpperCamelCase : "Conversation" ): '''simple docstring''' __UpperCAmelCase : Optional[int] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCamelCase , add_special_tokens=UpperCamelCase ) + [self.eos_token_id] ) if len(UpperCamelCase ) > self.model_max_length: __UpperCAmelCase : Any = input_ids[-self.model_max_length :] return input_ids
362
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase : Optional[int] = logging.get_logger(__name__) if is_vision_available(): import PIL class lowerCamelCase__ ( A ): """simple docstring""" __a = ["""pixel_values"""] def __init__( self : Tuple , UpperCamelCase : bool = True , UpperCamelCase : Dict[str, int] = None , UpperCamelCase : PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase : bool = True , UpperCamelCase : Dict[str, int] = None , UpperCamelCase : bool = True , UpperCamelCase : Union[int, float] = 1 / 255 , UpperCamelCase : bool = True , UpperCamelCase : Optional[Union[float, List[float]]] = None , UpperCamelCase : Optional[Union[float, List[float]]] = None , UpperCamelCase : bool = True , **UpperCamelCase : str , ): '''simple docstring''' super().__init__(**UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = size if size is not None else {"""shortest_edge""": 224} __UpperCAmelCase : str = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase ) __UpperCAmelCase : str = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} __UpperCAmelCase : str = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase , param_name="""crop_size""" ) __UpperCAmelCase : int = do_resize __UpperCAmelCase : Tuple = size __UpperCAmelCase : Optional[Any] = resample __UpperCAmelCase : Any = do_center_crop __UpperCAmelCase : int = crop_size __UpperCAmelCase : Optional[int] = do_rescale __UpperCAmelCase : List[Any] = rescale_factor __UpperCAmelCase : Tuple = do_normalize __UpperCAmelCase : Any = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __UpperCAmelCase : List[Any] = image_std if image_std is not None else OPENAI_CLIP_STD __UpperCAmelCase : List[Any] = do_convert_rgb def lowerCamelCase__ ( self : List[Any] , UpperCamelCase : np.ndarray , UpperCamelCase : Dict[str, int] , UpperCamelCase : PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase : List[Any] , ): '''simple docstring''' __UpperCAmelCase : Dict = get_size_dict(UpperCamelCase , default_to_square=UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) __UpperCAmelCase : int = get_resize_output_image_size(UpperCamelCase , size=size["""shortest_edge"""] , default_to_square=UpperCamelCase ) return resize(UpperCamelCase , size=UpperCamelCase , resample=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : List[Any] , UpperCamelCase : np.ndarray , UpperCamelCase : Dict[str, int] , UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase : Dict , ): '''simple docstring''' __UpperCAmelCase : Optional[int] = get_size_dict(UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(UpperCamelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : Any , UpperCamelCase : np.ndarray , UpperCamelCase : Union[int, float] , UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase : Any , ): '''simple docstring''' return rescale(UpperCamelCase , scale=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : Any , UpperCamelCase : np.ndarray , UpperCamelCase : Union[float, List[float]] , UpperCamelCase : Union[float, List[float]] , UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase : Any , ): '''simple docstring''' return normalize(UpperCamelCase , mean=UpperCamelCase , std=UpperCamelCase , data_format=UpperCamelCase , **UpperCamelCase ) def lowerCamelCase__ ( self : Tuple , UpperCamelCase : ImageInput , UpperCamelCase : bool = None , UpperCamelCase : Dict[str, int] = None , UpperCamelCase : PILImageResampling = None , UpperCamelCase : bool = None , UpperCamelCase : int = None , UpperCamelCase : bool = None , UpperCamelCase : float = None , UpperCamelCase : bool = None , UpperCamelCase : Optional[Union[float, List[float]]] = None , UpperCamelCase : Optional[Union[float, List[float]]] = None , UpperCamelCase : bool = None , UpperCamelCase : Optional[Union[str, TensorType]] = None , UpperCamelCase : Optional[ChannelDimension] = ChannelDimension.FIRST , **UpperCamelCase : Any , ): '''simple docstring''' __UpperCAmelCase : str = do_resize if do_resize is not None else self.do_resize __UpperCAmelCase : Dict = size if size is not None else self.size __UpperCAmelCase : Optional[Any] = get_size_dict(UpperCamelCase , param_name="""size""" , default_to_square=UpperCamelCase ) __UpperCAmelCase : Dict = resample if resample is not None else self.resample __UpperCAmelCase : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCAmelCase : str = crop_size if crop_size is not None else self.crop_size __UpperCAmelCase : Dict = get_size_dict(UpperCamelCase , param_name="""crop_size""" , default_to_square=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale __UpperCAmelCase : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCAmelCase : Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize __UpperCAmelCase : Any = image_mean if image_mean is not None else self.image_mean __UpperCAmelCase : Any = image_std if image_std is not None else self.image_std __UpperCAmelCase : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __UpperCAmelCase : List[str] = make_list_of_images(UpperCamelCase ) if not valid_images(UpperCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: __UpperCAmelCase : int = [convert_to_rgb(UpperCamelCase ) for image in images] # All transformations expect numpy arrays. __UpperCAmelCase : Tuple = [to_numpy_array(UpperCamelCase ) for image in images] if do_resize: __UpperCAmelCase : Optional[int] = [self.resize(image=UpperCamelCase , size=UpperCamelCase , resample=UpperCamelCase ) for image in images] if do_center_crop: __UpperCAmelCase : int = [self.center_crop(image=UpperCamelCase , size=UpperCamelCase ) for image in images] if do_rescale: __UpperCAmelCase : Dict = [self.rescale(image=UpperCamelCase , scale=UpperCamelCase ) for image in images] if do_normalize: __UpperCAmelCase : Optional[Any] = [self.normalize(image=UpperCamelCase , mean=UpperCamelCase , std=UpperCamelCase ) for image in images] __UpperCAmelCase : Any = [to_channel_dimension_format(UpperCamelCase , UpperCamelCase ) for image in images] __UpperCAmelCase : Any = {"""pixel_values""": images} return BatchFeature(data=UpperCamelCase , tensor_type=UpperCamelCase )
320
0
"""simple docstring""" from math import sqrt def UpperCAmelCase ( UpperCAmelCase = 1000000 ) -> int: snake_case_ = 0 snake_case_ = 0 snake_case_ = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(UpperCAmelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
69
from __future__ import annotations import bisect def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0 , _lowerCamelCase = -1 ): '''simple docstring''' if hi < 0: _lowerCAmelCase : int = len(_lowerCamelCase ) while lo < hi: _lowerCAmelCase : Optional[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: _lowerCAmelCase : Union[str, Any] = mid + 1 else: _lowerCAmelCase : str = mid return lo def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0 , _lowerCamelCase = -1 ): '''simple docstring''' if hi < 0: _lowerCAmelCase : str = len(_lowerCamelCase ) while lo < hi: _lowerCAmelCase : Tuple = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: _lowerCAmelCase : Dict = mid + 1 else: _lowerCAmelCase : str = mid return lo def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0 , _lowerCamelCase = -1 ): '''simple docstring''' sorted_collection.insert(bisect_left(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase ) def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0 , _lowerCamelCase = -1 ): '''simple docstring''' sorted_collection.insert(bisect_right(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase ) def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : int = 0 _lowerCAmelCase : Union[str, Any] = len(_lowerCamelCase ) - 1 while left <= right: _lowerCAmelCase : int = left + (right - left) // 2 _lowerCAmelCase : int = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: _lowerCAmelCase : str = midpoint - 1 else: _lowerCAmelCase : Any = midpoint + 1 return None def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Tuple = bisect.bisect_left(_lowerCamelCase , _lowerCamelCase ) if index != len(_lowerCamelCase ) and sorted_collection[index] == item: return index return None def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' if right < left: return None _lowerCAmelCase : Optional[int] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , midpoint - 1 ) else: return binary_search_by_recursion(_lowerCamelCase , _lowerCamelCase , midpoint + 1 , _lowerCamelCase ) if __name__ == "__main__": _snake_case = input("Enter numbers separated by comma:\n").strip() _snake_case = sorted(int(item) for item in user_input.split(",")) _snake_case = int(input("Enter a single number to be found in the list:\n")) _snake_case = binary_search(collection, target) if result is None: print(f'''{target} was not found in {collection}.''') else: print(f'''{target} was found at position {result} in {collection}.''')
36
0
from __future__ import annotations from collections.abc import MutableSequence class __lowerCamelCase : def __init__( self , lowerCamelCase , lowerCamelCase ) -> None: if len(lowerCamelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) snake_case_ = list(lowerCamelCase ) snake_case_ = degree def __add__( self , lowerCamelCase ) -> Polynomial: if self.degree > polynomial_a.degree: snake_case_ = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCamelCase ) else: snake_case_ = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCamelCase ) def __sub__( self , lowerCamelCase ) -> Polynomial: return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self ) -> Polynomial: return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self , lowerCamelCase ) -> Polynomial: snake_case_ = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCamelCase ) def lowerCAmelCase_ ( self , lowerCamelCase ) -> int | float: snake_case_ = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ) -> str: snake_case_ = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase ) return polynomial def __repr__( self ) -> str: return self.__str__() def lowerCAmelCase_ ( self ) -> Polynomial: snake_case_ = [0] * self.degree for i in range(self.degree ): snake_case_ = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCamelCase ) def lowerCAmelCase_ ( self , lowerCamelCase = 0 ) -> Polynomial: snake_case_ = [0] * (self.degree + 2) snake_case_ = constant for i in range(self.degree + 1 ): snake_case_ = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCamelCase ) def __eq__( self , lowerCamelCase ) -> bool: if not isinstance(lowerCamelCase , lowerCamelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self , lowerCamelCase ) -> bool: return not self.__eq__(lowerCamelCase )
34
import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase_ = re.compile(R'''\b(a|an|the)\b''', re.UNICODE) lowerCamelCase_ = None def UpperCamelCase( ) -> List[Any]: '''simple docstring''' snake_case_ = argparse.ArgumentParser("""Official evaluation script for SQuAD version 2.0.""" ) parser.add_argument("""data_file""" , metavar="""data.json""" , help="""Input data JSON file.""" ) parser.add_argument("""pred_file""" , metavar="""pred.json""" , help="""Model predictions.""" ) parser.add_argument( """--out-file""" , """-o""" , metavar="""eval.json""" , help="""Write accuracy metrics to file (default is stdout).""" ) parser.add_argument( """--na-prob-file""" , """-n""" , metavar="""na_prob.json""" , help="""Model estimates of probability of no answer.""" ) parser.add_argument( """--na-prob-thresh""" , """-t""" , type=lowercase_ , default=1.0 , help="""Predict \"\" if no-answer probability exceeds this (default = 1.0).""" , ) parser.add_argument( """--out-image-dir""" , """-p""" , metavar="""out_images""" , default=lowercase_ , help="""Save precision-recall curves to directory.""" ) parser.add_argument("""--verbose""" , """-v""" , action="""store_true""" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def UpperCamelCase( lowercase_ ) -> Union[str, Any]: '''simple docstring''' snake_case_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: snake_case_ = bool(qa["""answers"""]["""text"""] ) return qid_to_has_ans def UpperCamelCase( lowercase_ ) -> Tuple: '''simple docstring''' def remove_articles(lowercase_ ): return ARTICLES_REGEX.sub(""" """ , lowercase_ ) def white_space_fix(lowercase_ ): return " ".join(text.split() ) def remove_punc(lowercase_ ): snake_case_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def UpperCamelCase( lowercase_ ) -> Dict: '''simple docstring''' if not s: return [] return normalize_answer(lowercase_ ).split() def UpperCamelCase( lowercase_ , lowercase_ ) -> Dict: '''simple docstring''' return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def UpperCamelCase( lowercase_ , lowercase_ ) -> Any: '''simple docstring''' snake_case_ = get_tokens(lowercase_ ) snake_case_ = get_tokens(lowercase_ ) snake_case_ = collections.Counter(lowercase_ ) & collections.Counter(lowercase_ ) snake_case_ = sum(common.values() ) if len(lowercase_ ) == 0 or len(lowercase_ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 snake_case_ = 1.0 * num_same / len(lowercase_ ) snake_case_ = 1.0 * num_same / len(lowercase_ ) snake_case_ = (2 * precision * recall) / (precision + recall) return fa def UpperCamelCase( lowercase_ , lowercase_ ) -> List[str]: '''simple docstring''' snake_case_ = {} snake_case_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: snake_case_ = qa["""id"""] snake_case_ = [t for t in qa["""answers"""]["""text"""] if normalize_answer(lowercase_ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string snake_case_ = [""""""] if qid not in preds: print(f'''Missing prediction for {qid}''' ) continue snake_case_ = preds[qid] # Take max over all gold answers snake_case_ = max(compute_exact(lowercase_ , lowercase_ ) for a in gold_answers ) snake_case_ = max(compute_fa(lowercase_ , lowercase_ ) for a in gold_answers ) return exact_scores, fa_scores def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Union[str, Any]: '''simple docstring''' snake_case_ = {} for qid, s in scores.items(): snake_case_ = na_probs[qid] > na_prob_thresh if pred_na: snake_case_ = float(not qid_to_has_ans[qid] ) else: snake_case_ = s return new_scores def UpperCamelCase( lowercase_ , lowercase_ , lowercase_=None ) -> Dict: '''simple docstring''' if not qid_list: snake_case_ = len(lowercase_ ) return collections.OrderedDict( [ ("""exact""", 1_00.0 * sum(exact_scores.values() ) / total), ("""f1""", 1_00.0 * sum(fa_scores.values() ) / total), ("""total""", total), ] ) else: snake_case_ = len(lowercase_ ) return collections.OrderedDict( [ ("""exact""", 1_00.0 * sum(exact_scores[k] for k in qid_list ) / total), ("""f1""", 1_00.0 * sum(fa_scores[k] for k in qid_list ) / total), ("""total""", total), ] ) def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ ) -> Any: '''simple docstring''' for k in new_eval: snake_case_ = new_eval[k] def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Dict: '''simple docstring''' plt.step(lowercase_ , lowercase_ , color="""b""" , alpha=0.2 , where="""post""" ) plt.fill_between(lowercase_ , lowercase_ , step="""post""" , alpha=0.2 , color="""b""" ) plt.xlabel("""Recall""" ) plt.ylabel("""Precision""" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(lowercase_ ) plt.savefig(lowercase_ ) plt.clf() def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None ) -> Dict: '''simple docstring''' snake_case_ = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) snake_case_ = 0.0 snake_case_ = 1.0 snake_case_ = 0.0 snake_case_ = [1.0] snake_case_ = [0.0] snake_case_ = 0.0 for i, qid in enumerate(lowercase_ ): if qid_to_has_ans[qid]: true_pos += scores[qid] snake_case_ = true_pos / float(i + 1 ) snake_case_ = true_pos / float(lowercase_ ) if i == len(lowercase_ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(lowercase_ ) recalls.append(lowercase_ ) if out_image: plot_pr_curve(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) return {"ap": 1_00.0 * avg_prec} def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> str: '''simple docstring''' if out_image_dir and not os.path.exists(lowercase_ ): os.makedirs(lowercase_ ) snake_case_ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_exact.png""" ) , title="""Precision-Recall curve for Exact Match score""" , ) snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_f1.png""" ) , title="""Precision-Recall curve for F1 score""" , ) snake_case_ = {k: float(lowercase_ ) for k, v in qid_to_has_ans.items()} snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_oracle.png""" ) , title="""Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)""" , ) merge_eval(lowercase_ , lowercase_ , """pr_exact""" ) merge_eval(lowercase_ , lowercase_ , """pr_f1""" ) merge_eval(lowercase_ , lowercase_ , """pr_oracle""" ) def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> int: '''simple docstring''' if not qid_list: return snake_case_ = [na_probs[k] for k in qid_list] snake_case_ = np.ones_like(lowercase_ ) / float(len(lowercase_ ) ) plt.hist(lowercase_ , weights=lowercase_ , bins=20 , range=(0.0, 1.0) ) plt.xlabel("""Model probability of no-answer""" ) plt.ylabel("""Proportion of dataset""" ) plt.title(f'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(lowercase_ , f'''na_prob_hist_{name}.png''' ) ) plt.clf() def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Optional[int]: '''simple docstring''' snake_case_ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) snake_case_ = num_no_ans snake_case_ = cur_score snake_case_ = 0.0 snake_case_ = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) for i, qid in enumerate(lowercase_ ): if qid not in scores: continue if qid_to_has_ans[qid]: snake_case_ = scores[qid] else: if preds[qid]: snake_case_ = -1 else: snake_case_ = 0 cur_score += diff if cur_score > best_score: snake_case_ = cur_score snake_case_ = na_probs[qid] return 1_00.0 * best_score / len(lowercase_ ), best_thresh def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Any: '''simple docstring''' snake_case_ , snake_case_ = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) snake_case_ , snake_case_ = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) snake_case_ = best_exact snake_case_ = exact_thresh snake_case_ = best_fa snake_case_ = fa_thresh def UpperCamelCase( ) -> Union[str, Any]: '''simple docstring''' with open(OPTS.data_file ) as f: snake_case_ = json.load(lowercase_ ) snake_case_ = dataset_json["""data"""] with open(OPTS.pred_file ) as f: snake_case_ = json.load(lowercase_ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: snake_case_ = json.load(lowercase_ ) else: snake_case_ = {k: 0.0 for k in preds} snake_case_ = make_qid_to_has_ans(lowercase_ ) # maps qid to True/False snake_case_ = [k for k, v in qid_to_has_ans.items() if v] snake_case_ = [k for k, v in qid_to_has_ans.items() if not v] snake_case_ , snake_case_ = get_raw_scores(lowercase_ , lowercase_ ) snake_case_ = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) snake_case_ = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) snake_case_ = make_eval_dict(lowercase_ , lowercase_ ) if has_ans_qids: snake_case_ = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , """HasAns""" ) if no_ans_qids: snake_case_ = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , """NoAns""" ) if OPTS.na_prob_file: find_all_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , OPTS.out_image_dir ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , """hasAns""" ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , """noAns""" ) if OPTS.out_file: with open(OPTS.out_file , """w""" ) as f: json.dump(lowercase_ , lowercase_ ) else: print(json.dumps(lowercase_ , indent=2 ) ) if __name__ == "__main__": lowerCamelCase_ = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
34
1
import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class A ( __lowercase , unittest.TestCase ): __snake_case = FlaxAutoencoderKL @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = 4 lowerCAmelCase_ = 3 lowerCAmelCase_ = (32, 32) lowerCAmelCase_ = jax.random.PRNGKey(0 ) lowerCAmelCase_ = jax.random.uniform(_a, ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } lowerCAmelCase_ = self.dummy_input return init_dict, inputs_dict
278
from collections.abc import Callable class UpperCAmelCase_ : """simple docstring""" def __init__( self , _a = None ) -> None: # Stores actual heap items. _a : list = [] # Stores indexes of each item for supporting updates and deletion. _a : dict = {} # Stores current size of heap. _a : Tuple = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. _a : Dict = key or (lambda _a : x) def __lowercase ( self , _a ) -> int | None: return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self , _a ) -> int | None: _a : Optional[int] = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self , _a ) -> int | None: _a : int = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self , _a , _a ) -> None: _a , _a : Union[str, Any] = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. _a , _a : List[Any] = self.arr[j], self.arr[i] def __lowercase ( self , _a , _a ) -> bool: return self.arr[i][1] < self.arr[j][1] def __lowercase ( self , _a ) -> int: _a : Dict = self._left(_a ) _a : str = self._right(_a ) _a : str = i if left is not None and not self._cmp(_a , _a ): _a : Optional[Any] = left if right is not None and not self._cmp(_a , _a ): _a : Any = right return valid_parent def __lowercase ( self , _a ) -> None: _a : List[str] = self._parent(_a ) while parent is not None and not self._cmp(_a , _a ): self._swap(_a , _a ) _a , _a : Any = parent, self._parent(_a ) def __lowercase ( self , _a ) -> None: _a : List[Any] = self._get_valid_parent(_a ) while valid_parent != index: self._swap(_a , _a ) _a , _a : int = valid_parent, self._get_valid_parent(_a ) def __lowercase ( self , _a , _a ) -> None: if item not in self.pos_map: return _a : str = self.pos_map[item] _a : List[Any] = [item, self.key(_a )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(_a ) self._heapify_down(_a ) def __lowercase ( self , _a ) -> None: if item not in self.pos_map: return _a : Tuple = self.pos_map[item] del self.pos_map[item] _a : Tuple = self.arr[self.size - 1] _a : str = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(_a ) self._heapify_down(_a ) def __lowercase ( self , _a , _a ) -> None: _a : Union[str, Any] = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(_a )] ) else: _a : Optional[int] = [item, self.key(_a )] _a : Tuple = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self ) -> tuple | None: return self.arr[0] if self.size else None def __lowercase ( self ) -> tuple | None: _a : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def __UpperCAmelCase ( ) -> None: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
235
0
"""simple docstring""" def __lowerCamelCase ( a_ : int , a_ : int ) -> Any: if b == 0: return 1 if (b % 2) == 0: return actual_power(__a , int(b / 2 ) ) * actual_power(__a , int(b / 2 ) ) else: return a * actual_power(__a , int(b / 2 ) ) * actual_power(__a , int(b / 2 ) ) def __lowerCamelCase ( a_ : int , a_ : int ) -> Any: if b < 0: return 1 / actual_power(__a , __a ) return actual_power(__a , __a ) if __name__ == "__main__": print(power(-2, -3))
355
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "google/vivit-b-16x2-kinetics400": ( "https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json" ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = '''vivit''' def __init__( self ,SCREAMING_SNAKE_CASE__=2_24 ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=[2, 16, 16] ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu_fast" ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-06 ,SCREAMING_SNAKE_CASE__=True ,**SCREAMING_SNAKE_CASE__ ,) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_size __SCREAMING_SNAKE_CASE :List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE :List[str] = num_attention_heads __SCREAMING_SNAKE_CASE :Any = intermediate_size __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_act __SCREAMING_SNAKE_CASE :int = hidden_dropout_prob __SCREAMING_SNAKE_CASE :Union[str, Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Dict = initializer_range __SCREAMING_SNAKE_CASE :Dict = layer_norm_eps __SCREAMING_SNAKE_CASE :Any = image_size __SCREAMING_SNAKE_CASE :Any = num_frames __SCREAMING_SNAKE_CASE :Any = tubelet_size __SCREAMING_SNAKE_CASE :Tuple = num_channels __SCREAMING_SNAKE_CASE :List[str] = qkv_bias super().__init__(**SCREAMING_SNAKE_CASE__ )
239
0
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''deepmind/language-perceiver''': '''https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json''', # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = '''perceiver''' def __init__( self : Any , __UpperCAmelCase : Optional[Any]=256 , __UpperCAmelCase : str=1280 , __UpperCAmelCase : List[str]=768 , __UpperCAmelCase : Union[str, Any]=1 , __UpperCAmelCase : List[Any]=26 , __UpperCAmelCase : int=8 , __UpperCAmelCase : Union[str, Any]=8 , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : Optional[Any]="kv" , __UpperCAmelCase : int=1 , __UpperCAmelCase : int=1 , __UpperCAmelCase : str="gelu" , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : int=0.02 , __UpperCAmelCase : Optional[Any]=1E-12 , __UpperCAmelCase : str=True , __UpperCAmelCase : Dict=262 , __UpperCAmelCase : List[str]=2048 , __UpperCAmelCase : List[Any]=56 , __UpperCAmelCase : Dict=[368, 496] , __UpperCAmelCase : Union[str, Any]=16 , __UpperCAmelCase : Optional[Any]=1920 , __UpperCAmelCase : Any=16 , __UpperCAmelCase : Any=[1, 16, 224, 224] , **__UpperCAmelCase : Optional[int] , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) _A = num_latents _A = d_latents _A = d_model _A = num_blocks _A = num_self_attends_per_block _A = num_self_attention_heads _A = num_cross_attention_heads _A = qk_channels _A = v_channels _A = cross_attention_shape_for_attention _A = self_attention_widening_factor _A = cross_attention_widening_factor _A = hidden_act _A = attention_probs_dropout_prob _A = initializer_range _A = layer_norm_eps _A = use_query_residual # masked language modeling attributes _A = vocab_size _A = max_position_embeddings # image classification attributes _A = image_size # flow attributes _A = train_size # multimodal autoencoding attributes _A = num_frames _A = audio_samples_per_frame _A = samples_per_patch _A = output_shape class _UpperCAmelCase ( snake_case_ ): """simple docstring""" @property def lowerCAmelCase ( self : Tuple ): '''simple docstring''' if self.task == "multiple-choice": _A = {0: "batch", 1: "choice", 2: "sequence"} else: _A = {0: "batch", 1: "sequence"} return OrderedDict( [ ("inputs", dynamic_axis), ("attention_mask", dynamic_axis), ] ) @property def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' return 1E-4 def lowerCAmelCase ( self : Dict , __UpperCAmelCase : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , __UpperCAmelCase : int = -1 , __UpperCAmelCase : int = -1 , __UpperCAmelCase : int = -1 , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[TensorType] = None , __UpperCAmelCase : int = 3 , __UpperCAmelCase : int = 40 , __UpperCAmelCase : int = 40 , ): '''simple docstring''' if isinstance(__UpperCAmelCase , __UpperCAmelCase ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _A = compute_effective_axis_dimension( __UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _A = preprocessor.num_special_tokens_to_add(__UpperCAmelCase ) _A = compute_effective_axis_dimension( __UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__UpperCAmelCase ) # Generate dummy inputs according to compute batch and sequence _A = [" ".join(["a"] ) * seq_length] * batch_size _A = dict(preprocessor(__UpperCAmelCase , return_tensors=__UpperCAmelCase ) ) _A = inputs.pop("input_ids" ) return inputs elif isinstance(__UpperCAmelCase , __UpperCAmelCase ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _A = compute_effective_axis_dimension(__UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch ) _A = self._generate_dummy_images(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _A = dict(preprocessor(images=__UpperCAmelCase , return_tensors=__UpperCAmelCase ) ) _A = inputs.pop("pixel_values" ) return inputs else: raise ValueError( "Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor." )
79
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase: Dict = { """configuration_whisper""": ["""WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """WhisperConfig""", """WhisperOnnxConfig"""], """feature_extraction_whisper""": ["""WhisperFeatureExtractor"""], """processing_whisper""": ["""WhisperProcessor"""], """tokenization_whisper""": ["""WhisperTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase: Dict = ["""WhisperTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase: int = [ """WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """WhisperForConditionalGeneration""", """WhisperModel""", """WhisperPreTrainedModel""", """WhisperForAudioClassification""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase: Tuple = [ """TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFWhisperForConditionalGeneration""", """TFWhisperModel""", """TFWhisperPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase: Optional[Any] = [ """FlaxWhisperForConditionalGeneration""", """FlaxWhisperModel""", """FlaxWhisperPreTrainedModel""", """FlaxWhisperForAudioClassification""", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys UpperCAmelCase: Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
366
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class UpperCamelCase ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self ): _lowercase : str = tempfile.mkdtemp() # fmt: off _lowercase : List[Any] = ["""l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on _lowercase : Optional[int] = dict(zip(UpperCAmelCase_ ,range(len(UpperCAmelCase_ ) ) ) ) _lowercase : Optional[int] = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] _lowercase : Optional[int] = {"""unk_token""": """<unk>"""} _lowercase : List[str] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) _lowercase : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(UpperCAmelCase_ ) ) _lowercase : Dict = { """do_resize""": True, """size""": 20, """do_center_crop""": True, """crop_size""": 18, """do_normalize""": True, """image_mean""": [0.48145466, 0.4578275, 0.40821073], """image_std""": [0.26862954, 0.26130258, 0.27577711], } _lowercase : List[Any] = os.path.join(self.tmpdirname ,UpperCAmelCase_ ) with open(self.image_processor_file ,"""w""" ,encoding="""utf-8""" ) as fp: json.dump(UpperCAmelCase_ ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ,**UpperCAmelCase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,**UpperCAmelCase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,**UpperCAmelCase_ ): return CLIPImageProcessor.from_pretrained(self.tmpdirname ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = [np.random.randint(2_55 ,size=(3, 30, 4_00) ,dtype=np.uinta )] _lowercase : Tuple = [Image.fromarray(np.moveaxis(UpperCAmelCase_ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = self.get_tokenizer() _lowercase : List[Any] = self.get_rust_tokenizer() _lowercase : List[Any] = self.get_image_processor() _lowercase : List[str] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) processor_slow.save_pretrained(self.tmpdirname ) _lowercase : Optional[int] = CLIPProcessor.from_pretrained(self.tmpdirname ,use_fast=UpperCAmelCase_ ) _lowercase : Union[str, Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) processor_fast.save_pretrained(self.tmpdirname ) _lowercase : List[str] = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() ,tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() ,tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() ,tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer ,UpperCAmelCase_ ) self.assertIsInstance(processor_fast.tokenizer ,UpperCAmelCase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() ,image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() ,image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor ,UpperCAmelCase_ ) self.assertIsInstance(processor_fast.image_processor ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : str = CLIPProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _lowercase : Union[str, Any] = self.get_tokenizer(bos_token="""(BOS)""" ,eos_token="""(EOS)""" ) _lowercase : Optional[int] = self.get_image_processor(do_normalize=UpperCAmelCase_ ,padding_value=1.0 ) _lowercase : int = CLIPProcessor.from_pretrained( self.tmpdirname ,bos_token="""(BOS)""" ,eos_token="""(EOS)""" ,do_normalize=UpperCAmelCase_ ,padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,UpperCAmelCase_ ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = self.get_image_processor() _lowercase : Optional[int] = self.get_tokenizer() _lowercase : Union[str, Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) _lowercase : int = self.prepare_image_inputs() _lowercase : str = image_processor(UpperCAmelCase_ ,return_tensors="""np""" ) _lowercase : int = processor(images=UpperCAmelCase_ ,return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() ,input_processor[key].sum() ,delta=1E-2 ) def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = self.get_image_processor() _lowercase : Optional[Any] = self.get_tokenizer() _lowercase : Union[str, Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) _lowercase : List[Any] = """lower newer""" _lowercase : Any = processor(text=UpperCAmelCase_ ) _lowercase : Union[str, Any] = tokenizer(UpperCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = self.get_image_processor() _lowercase : List[Any] = self.get_tokenizer() _lowercase : List[Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) _lowercase : str = """lower newer""" _lowercase : List[Any] = self.prepare_image_inputs() _lowercase : Optional[int] = processor(text=UpperCAmelCase_ ,images=UpperCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) ,["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase_ ): processor() def lowerCamelCase__ ( self ): _lowercase : Dict = self.get_image_processor() _lowercase : List[Any] = self.get_tokenizer() _lowercase : Union[str, Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) _lowercase : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _lowercase : int = processor.batch_decode(UpperCAmelCase_ ) _lowercase : Tuple = tokenizer.batch_decode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = self.get_image_processor() _lowercase : List[Any] = self.get_tokenizer() _lowercase : List[Any] = CLIPProcessor(tokenizer=UpperCAmelCase_ ,image_processor=UpperCAmelCase_ ) _lowercase : Optional[Any] = """lower newer""" _lowercase : Any = self.prepare_image_inputs() _lowercase : Optional[int] = processor(text=UpperCAmelCase_ ,images=UpperCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) ,processor.model_input_names )
336
0
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class A ( ctypes.Structure ): """simple docstring""" lowerCamelCase = [('size', ctypes.c_int), ('visible', ctypes.c_byte)] def _snake_case( ) -> Optional[int]: '''simple docstring''' if os.name == "nt": A__ = CursorInfo() A__ = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE__ , ctypes.byref(SCREAMING_SNAKE_CASE__ ) ) A__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE__ , ctypes.byref(SCREAMING_SNAKE_CASE__ ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def _snake_case( ) -> Optional[int]: '''simple docstring''' if os.name == "nt": A__ = CursorInfo() A__ = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE__ , ctypes.byref(SCREAMING_SNAKE_CASE__ ) ) A__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE__ , ctypes.byref(SCREAMING_SNAKE_CASE__ ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def _snake_case( ) -> Union[str, Any]: '''simple docstring''' try: hide_cursor() yield finally: show_cursor()
7
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowerCAmelCase_ ( _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Dict = "mock-s3-bucket" __magic_name__ : Any = F'''s3://{mock_bucket}''' __magic_name__ : str = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("s3://" ) is False __magic_name__ : Tuple = "./local/path" __magic_name__ : Optional[Any] = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def lowerCAmelCase_ ( _snake_case : List[str] ) -> Optional[Any]: '''simple docstring''' __magic_name__ : str = is_remote_filesystem(_snake_case ) assert is_remote is True __magic_name__ : Optional[int] = fsspec.filesystem("file" ) __magic_name__ : int = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Tuple , _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Any ) -> int: '''simple docstring''' __magic_name__ : Any = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __magic_name__ : str = input_paths[compression_fs_class.protocol] if input_path is None: __magic_name__ : Dict = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) __magic_name__ : str = fsspec.filesystem(compression_fs_class.protocol , fo=_snake_case ) assert isinstance(_snake_case , _snake_case ) __magic_name__ : int = os.path.basename(_snake_case ) __magic_name__ : Optional[int] = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(_snake_case , "r" , encoding="utf-8" ) as f, open(_snake_case , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ) -> str: '''simple docstring''' __magic_name__ : int = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __magic_name__ : int = compressed_file_paths[protocol] __magic_name__ : Tuple = "dataset.jsonl" __magic_name__ : List[str] = F'''{protocol}://{member_file_path}::{compressed_file_path}''' __magic_name__ , *__magic_name__ : Optional[Any] = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : List[str] , _snake_case : Tuple ) -> str: '''simple docstring''' __magic_name__ : int = hf_api.dataset_info(_snake_case , token=_snake_case ) __magic_name__ : Optional[Any] = HfFileSystem(repo_info=_snake_case , token=_snake_case ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(_snake_case ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def lowerCAmelCase_ ( ) -> Optional[int]: '''simple docstring''' __magic_name__ : Optional[Any] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case , _snake_case , clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
281
0
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. SCREAMING_SNAKE_CASE_ = abspath(join(dirname(dirname(dirname(__file__))), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(_SCREAMING_SNAKE_CASE ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(_SCREAMING_SNAKE_CASE , id=_SCREAMING_SNAKE_CASE )
193
import unittest from dataclasses import dataclass import pytest from accelerate.commands.config.config_args import SageMakerConfig from accelerate.utils import ComputeEnvironment from accelerate.utils.launch import _convert_nargs_to_dict @dataclass class UpperCamelCase__ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case : int = ComputeEnvironment.AMAZON_SAGEMAKER __snake_case : List[Any] = True __snake_case : Optional[int] = "ml.p3.2xlarge" __snake_case : List[str] = "accelerate_sagemaker_execution_role" __snake_case : Tuple = "hf-sm" __snake_case : Any = "us-east-1" __snake_case : Union[str, Any] = 1 __snake_case : Dict = "accelerate-sagemaker-1" __snake_case : Tuple = "1.6" __snake_case : List[str] = "4.4" __snake_case : str = "train.py" __snake_case : List[str] = [ "--model_name_or_path", "bert", "--do_train", "False", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] __snake_case : Optional[int] = [ "--model_name_or_path", "bert", "--do_train", "--do_test", "False", "--do_predict", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = _convert_nargs_to_dict(MockLaunchConfig.success_training_script_args ) assert isinstance(converted_args["""model_name_or_path"""] ,lowerCamelCase__ ) assert isinstance(converted_args["""do_train"""] ,lowerCamelCase__ ) assert isinstance(converted_args["""epochs"""] ,lowerCamelCase__ ) assert isinstance(converted_args["""learning_rate"""] ,lowerCamelCase__ ) assert isinstance(converted_args["""max_steps"""] ,lowerCamelCase__ ) with pytest.raises(lowerCamelCase__ ): _convert_nargs_to_dict(MockLaunchConfig.fail_training_script_args )
193
1
def lowercase_ ( _A : str ): """simple docstring""" if not all(x.isalpha() for x in string ): raise ValueError("String must only contain alphabetic characters." ) lowerCamelCase__ : Any = sorted(string.lower() ) return len(__lowerCAmelCase ) == len(set(__lowerCAmelCase ) ) if __name__ == "__main__": A : Optional[Any] = input("Enter a string ").strip() A : Dict = is_isogram(input_str) print(f'{input_str} is {"an" if isogram else "not an"} isogram.')
184
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE__ : Optional[Any] = {"configuration_wavlm": ["WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "WavLMConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : int = [ "WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST", "WavLMForAudioFrameClassification", "WavLMForCTC", "WavLMForSequenceClassification", "WavLMForXVector", "WavLMModel", "WavLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
270
0
'''simple docstring''' import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets lowerCAmelCase_ : Optional[Any] = datasets.logging.get_logger(__name__) lowerCAmelCase_ : Optional[int] = '\\n@inproceedings{bleurt,\n title={BLEURT: Learning Robust Metrics for Text Generation},\n author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh},\n booktitle={ACL},\n year={2020},\n url={https://arxiv.org/abs/2004.04696}\n}\n' lowerCAmelCase_ : Any = '\\nBLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018)\nand then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune\nit for your specific application (the latter is expected to perform better).\n\nSee the project\'s README at https://github.com/google-research/bleurt#readme for more information.\n' lowerCAmelCase_ : Dict = '\nBLEURT score.\n\nArgs:\n `predictions` (list of str): prediction/candidate sentences\n `references` (list of str): reference sentences\n `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None.\n\nReturns:\n \'scores\': List of scores.\nExamples:\n\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> bleurt = datasets.load_metric("bleurt")\n >>> results = bleurt.compute(predictions=predictions, references=references)\n >>> print([round(v, 2) for v in results["scores"]])\n [1.03, 1.04]\n' lowerCAmelCase_ : Union[str, Any] = { 'bleurt-tiny-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip', 'bleurt-tiny-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip', 'bleurt-base-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip', 'bleurt-base-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip', 'bleurt-large-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip', 'bleurt-large-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip', 'BLEURT-20-D3': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip', 'BLEURT-20-D6': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip', 'BLEURT-20-D12': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip', 'BLEURT-20': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE (datasets.Metric ): """simple docstring""" def UpperCamelCase__ ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/google-research/bleurt" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/bleurt"] , reference_urls=["https://github.com/google-research/bleurt", "https://arxiv.org/abs/2004.04696"] , ) def UpperCamelCase__ ( self : List[Any] , __a : Optional[Any] ): # check that config name specifies a valid BLEURT model if self.config_name == "default": logger.warning( "Using default BLEURT-Base checkpoint for sequence maximum length 128. " "You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512')." ) _a = "bleurt-base-128" if self.config_name.lower() in CHECKPOINT_URLS: _a = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: _a = self.config_name.upper() else: raise KeyError( f'{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}' ) # download the model checkpoint specified by self.config_name and set up the scorer _a = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) _a = score.BleurtScorer(os.path.join(__a , __a ) ) def UpperCamelCase__ ( self : str , __a : int , __a : str ): _a = self.scorer.score(references=__a , candidates=__a ) return {"scores": scores}
346
'''simple docstring''' from random import randint, random def _lowerCamelCase ( lowercase : int , lowercase : int , lowercase : int , lowercase : bool = False , lowercase : bool = False , lowercase : int = 5 , ) -> list: _a = [[-1] * number_of_cells] # Create a highway without any car _a = 0 _a = max(lowercase , 0 ) while i < number_of_cells: _a = ( randint(0 , lowercase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def _lowerCamelCase ( lowercase : list , lowercase : int ) -> int: _a = 0 _a = highway_now[car_index + 1 :] for cell in range(len(lowercase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(lowercase , -1 ) def _lowerCamelCase ( lowercase : list , lowercase : float , lowercase : int ) -> list: _a = len(lowercase ) # Beforce calculations, the highway is empty _a = [-1] * number_of_cells for car_index in range(lowercase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed _a = min(highway_now[car_index] + 1 , lowercase ) # Number of empty cell before the next car _a = get_distance(lowercase , lowercase ) - 1 # We can't have the car causing an accident _a = min(next_highway[car_index] , lowercase ) if random() < probability: # Randomly, a driver will slow down _a = max(next_highway[car_index] - 1 , 0 ) return next_highway def _lowerCamelCase ( lowercase : list , lowercase : int , lowercase : float , lowercase : int ) -> list: _a = len(highway[0] ) for i in range(lowercase ): _a = update(highway[i] , lowercase , lowercase ) _a = [-1] * number_of_cells for car_index in range(lowercase ): _a = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) _a = (car_index + speed) % number_of_cells # Commit the change of position _a = speed highway.append(lowercase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
346
1
import os __lowerCAmelCase : int = {'I': 1, 'V': 5, 'X': 10, 'L': 50, 'C': 100, 'D': 500, 'M': 1000} def a__ ( A_ ): '''simple docstring''' __magic_name__ = 0 __magic_name__ = 0 while index < len(A_ ) - 1: __magic_name__ = SYMBOLS[numerals[index]] __magic_name__ = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def a__ ( A_ ): '''simple docstring''' __magic_name__ = """""" __magic_name__ = num // 1000 numerals += m_count * "M" num %= 1000 __magic_name__ = num // 100 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 100 __magic_name__ = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def a__ ( A_ = "/p089_roman.txt" ): '''simple docstring''' __magic_name__ = 0 with open(os.path.dirname(A_ ) + roman_numerals_filename ) as filea: __magic_name__ = filea.readlines() for line in lines: __magic_name__ = line.strip() __magic_name__ = parse_roman_numerals(A_ ) __magic_name__ = generate_roman_numerals(A_ ) savings += len(A_ ) - len(A_ ) return savings if __name__ == "__main__": print(F'''{solution() = }''')
88
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self : int , A : Tuple , A : int=3 , A : List[str]=32 , A : Dict=3 , A : Any=10 , A : Dict=[10, 20, 30, 40] , A : Optional[Any]=[1, 1, 2, 1] , A : Union[str, Any]=True , A : Optional[Any]=True , A : Any="relu" , A : Optional[Any]=3 , A : Tuple=None , ) -> Dict: lowercase_ : str = parent lowercase_ : List[Any] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : int = num_channels lowercase_ : int = embeddings_size lowercase_ : str = hidden_sizes lowercase_ : List[str] = depths lowercase_ : Dict = is_training lowercase_ : int = use_labels lowercase_ : Any = hidden_act lowercase_ : List[Any] = num_labels lowercase_ : Tuple = scope lowercase_ : Optional[Any] = len(A ) def A ( self : str ) -> Tuple: lowercase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_labels ) lowercase_ : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Dict ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def A ( self : str , A : Tuple , A : str , A : str ) -> str: lowercase_ : str = TFResNetModel(config=A ) lowercase_ : Union[str, Any] = model(A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Any , A : int , A : List[Any] , A : Optional[Any] ) -> Optional[Any]: lowercase_ : Tuple = self.num_labels lowercase_ : Union[str, Any] = TFResNetForImageClassification(A ) lowercase_ : Tuple = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : Dict = config_and_inputs lowercase_ : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Any = False def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : int = TFResNetModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Dict ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Dict ) -> List[Any]: return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def A ( self : Any ) -> Any: pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def A ( self : List[str] ) -> Optional[Any]: pass def A ( self : str ) -> Tuple: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(A ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : str = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : List[str] ) -> Tuple: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : List[Any] ) -> List[str]: def check_hidden_states_output(A : Union[str, Any] , A : int , A : List[Any] ): lowercase_ : int = model_class(A ) lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Any = self.model_tester.num_stages self.assertEqual(len(A ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase_ : List[str] = layer_type lowercase_ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : Optional[Any] = True check_hidden_states_output(A , A , A ) def A ( self : Optional[int] ) -> Tuple: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def A ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFResNetModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase ( ): lowercase_ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Any ) -> Optional[int]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A ( self : Any ) -> Optional[int]: lowercase_ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase_ : List[Any] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : List[str] = image_processor(images=A , return_tensors='''tf''' ) # forward pass lowercase_ : Tuple = model(**A ) # verify the logits lowercase_ : Optional[int] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Optional[Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A , atol=1e-4 ) )
33
0
"""simple docstring""" import math lowerCAmelCase__ = 10 lowerCAmelCase__ = 7 lowerCAmelCase__ = BALLS_PER_COLOUR * NUM_COLOURS def snake_case_ ( A_ : int = 20 ): '''simple docstring''' _lowerCamelCase : Tuple = math.comb(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) _lowerCamelCase : Optional[int] = math.comb(NUM_BALLS - BALLS_PER_COLOUR, SCREAMING_SNAKE_CASE_ ) _lowerCamelCase : List[str] = NUM_COLOURS * (1 - missing_colour / total) return F'''{result:.9f}''' if __name__ == "__main__": print(solution(20))
364
"""simple docstring""" from itertools import permutations def snake_case_ ( A_ : tuple ): '''simple docstring''' if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False _lowerCamelCase : Any = [7, 11, 13, 17] for i, test in enumerate(A_ ): if (num[i + 4] * 1_00 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def snake_case_ ( A_ : int = 10 ): '''simple docstring''' return sum( int(''''''.join(map(A_, A_ ) ) ) for num in permutations(range(A_ ) ) if is_substring_divisible(A_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
175
0
"""simple docstring""" import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCAmelCase :Optional[Any] = 16 _lowerCAmelCase :Optional[int] = 32 def lowerCamelCase_ (UpperCamelCase__ : Accelerator , UpperCamelCase__ : int = 16 ): _UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) _UpperCAmelCase : int = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(UpperCamelCase__ : str ): # max_length=None => use the model max length (it's actually the default) _UpperCAmelCase : int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _UpperCAmelCase : List[Any] = datasets.map( _lowerCAmelCase , batched=_lowerCAmelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _UpperCAmelCase : int = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(UpperCamelCase__ : Optional[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. _UpperCAmelCase : int = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _UpperCAmelCase : List[str] = 16 elif accelerator.mixed_precision != "no": _UpperCAmelCase : Optional[Any] = 8 else: _UpperCAmelCase : Optional[Any] = None return tokenizer.pad( _lowerCAmelCase , padding='''longest''' , max_length=_lowerCAmelCase , pad_to_multiple_of=_lowerCAmelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. _UpperCAmelCase : Optional[Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase , drop_last=_lowerCAmelCase ) _UpperCAmelCase : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase , drop_last=(accelerator.mixed_precision == '''fp8''') , ) return train_dataloader, eval_dataloader def lowerCamelCase_ (UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Tuple ): _UpperCAmelCase : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _UpperCAmelCase : Union[str, Any] = config['''lr'''] _UpperCAmelCase : List[str] = int(config['''num_epochs'''] ) _UpperCAmelCase : Union[str, Any] = int(config['''seed'''] ) _UpperCAmelCase : int = int(config['''batch_size'''] ) _UpperCAmelCase : Optional[Any] = evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation _UpperCAmelCase : Optional[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _UpperCAmelCase : List[str] = batch_size // MAX_GPU_BATCH_SIZE _UpperCAmelCase : Optional[Any] = MAX_GPU_BATCH_SIZE set_seed(_lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Tuple = get_dataloaders(_lowerCAmelCase , _lowerCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _UpperCAmelCase : Dict = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=_lowerCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _UpperCAmelCase : Union[str, Any] = model.to(accelerator.device ) # Instantiate optimizer _UpperCAmelCase : Tuple = AdamW(params=model.parameters() , lr=_lowerCAmelCase ) # Instantiate scheduler _UpperCAmelCase : int = get_linear_schedule_with_warmup( optimizer=_lowerCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_lowerCAmelCase ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : str = accelerator.prepare( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Now we train the model for epoch in range(_lowerCAmelCase ): model.train() for step, batch in enumerate(_lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _UpperCAmelCase : Union[str, Any] = model(**_lowerCAmelCase ) _UpperCAmelCase : str = outputs.loss _UpperCAmelCase : Optional[int] = loss / gradient_accumulation_steps accelerator.backward(_lowerCAmelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _UpperCAmelCase : int = model(**_lowerCAmelCase ) _UpperCAmelCase : Optional[int] = outputs.logits.argmax(dim=-1 ) _UpperCAmelCase , _UpperCAmelCase : Any = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=_lowerCAmelCase , references=_lowerCAmelCase , ) _UpperCAmelCase : List[str] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' , _lowerCAmelCase ) def lowerCamelCase_ (): _UpperCAmelCase : str = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=_lowerCAmelCase , default=_lowerCAmelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) _UpperCAmelCase : Any = parser.parse_args() _UpperCAmelCase : List[str] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": main()
263
"""simple docstring""" import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor __snake_case = logging.get_logger(__name__) class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
320
0
'''simple docstring''' from datetime import datetime import requests def lowercase_ ( lowerCAmelCase__ : str ): """simple docstring""" __UpperCAmelCase : List[str] = """https://downloadgram.net/wp-json/wppress/video-downloader/video?url=""" __UpperCAmelCase : Tuple = requests.get(base_url + url ).json()[0]["""urls"""][0]["""src"""] return requests.get(lowerCAmelCase__ ).content if __name__ == "__main__": _UpperCamelCase = input('''Enter Video/IGTV url: ''').strip() _UpperCamelCase = F'{datetime.now():%Y-%m-%d_%H:%M:%S}.mp4' with open(file_name, '''wb''') as fp: fp.write(download_video(url)) print(F'Done. Video saved to disk as {file_name}.')
16
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {'''vocab_file''': '''vocab.txt'''} _UpperCamelCase = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } _UpperCamelCase = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } _UpperCamelCase = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _A ( __SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Any = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : List[Any] = PRETRAINED_INIT_CONFIGURATION _SCREAMING_SNAKE_CASE : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : List[Any] = ConvBertTokenizer def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase="[UNK]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[PAD]" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase=True , __UpperCAmelCase=None , **__UpperCAmelCase , ) -> Optional[Any]: '''simple docstring''' super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , tokenize_chinese_chars=__UpperCAmelCase , strip_accents=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCAmelCase : Optional[int] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , __UpperCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , __UpperCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , __UpperCAmelCase ) != tokenize_chinese_chars ): __UpperCAmelCase : Dict = getattr(__UpperCAmelCase , normalizer_state.pop("""type""" ) ) __UpperCAmelCase : Union[str, Any] = do_lower_case __UpperCAmelCase : str = strip_accents __UpperCAmelCase : Union[str, Any] = tokenize_chinese_chars __UpperCAmelCase : List[Any] = normalizer_class(**__UpperCAmelCase ) __UpperCAmelCase : List[Any] = do_lower_case def __A ( self , __UpperCAmelCase , __UpperCAmelCase=None ) -> List[str]: '''simple docstring''' __UpperCAmelCase : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __A ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> List[int]: '''simple docstring''' __UpperCAmelCase : Optional[int] = [self.sep_token_id] __UpperCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __A ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> Tuple[str]: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase )
16
1
'''simple docstring''' import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _a ( __a ): def __init__( self : Optional[int] , *lowercase : Optional[Any] , lowercase : Optional[Any]=None , lowercase : Tuple=None , **lowercase : Dict ): '''simple docstring''' super().__init__(*lowercase , **lowercase ) UpperCAmelCase = eval_examples UpperCAmelCase = post_process_function def A ( self : str , lowercase : List[Any]=None , lowercase : Tuple=None , lowercase : Optional[Any]=None , lowercase : str = "eval" ): '''simple docstring''' UpperCAmelCase = self.eval_dataset if eval_dataset is None else eval_dataset UpperCAmelCase = self.get_eval_dataloader(lowercase ) UpperCAmelCase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop UpperCAmelCase = time.time() try: UpperCAmelCase = eval_loop( lowercase , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: UpperCAmelCase = compute_metrics UpperCAmelCase = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default UpperCAmelCase = self.post_process_function(lowercase , lowercase , output.predictions ) UpperCAmelCase = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): UpperCAmelCase = metrics.pop(lowercase ) metrics.update(output.metrics ) else: UpperCAmelCase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowercase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) UpperCAmelCase = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowercase ) return metrics def A ( self : Optional[Any] , lowercase : Optional[Any] , lowercase : Tuple , lowercase : Optional[Any]=None , lowercase : str = "test" ): '''simple docstring''' UpperCAmelCase = self.get_test_dataloader(lowercase ) # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop UpperCAmelCase = time.time() try: UpperCAmelCase = eval_loop( lowercase , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: UpperCAmelCase = compute_metrics UpperCAmelCase = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output UpperCAmelCase = self.post_process_function(lowercase , lowercase , output.predictions , '''predict''' ) UpperCAmelCase = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): UpperCAmelCase = metrics.pop(lowercase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowercase )
34
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def snake_case_ (_a : dict , _a : str , _a : set , _a : set , _a : dict , _a : dict , _a : PriorityQueue , _a : dict , _a : float | int , ): for nxt, d in graph[v]: if nxt in visited_forward: continue UpperCAmelCase = cst_fwd.get(_a , np.inf ) UpperCAmelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) UpperCAmelCase = new_cost_f UpperCAmelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: UpperCAmelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def snake_case_ (_a : str , _a : str , _a : dict , _a : dict ): UpperCAmelCase = -1 UpperCAmelCase = set() UpperCAmelCase = set() UpperCAmelCase = {source: 0} UpperCAmelCase = {destination: 0} UpperCAmelCase = {source: None} UpperCAmelCase = {destination: None} UpperCAmelCase = PriorityQueue() UpperCAmelCase = PriorityQueue() UpperCAmelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): UpperCAmelCase , UpperCAmelCase = queue_forward.get() visited_forward.add(_a ) UpperCAmelCase , UpperCAmelCase = queue_backward.get() visited_backward.add(_a ) UpperCAmelCase = pass_and_relaxation( _a , _a , _a , _a , _a , _a , _a , _a , _a , ) UpperCAmelCase = pass_and_relaxation( _a , _a , _a , _a , _a , _a , _a , _a , _a , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: UpperCAmelCase = shortest_distance return shortest_path_distance A ={ 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } A ={ 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
34
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCAmelCase : str = { """configuration_convnext""": ["""CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvNextConfig""", """ConvNextOnnxConfig"""] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = ["""ConvNextFeatureExtractor"""] _lowerCAmelCase : List[Any] = ["""ConvNextImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : int = [ """CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvNextForImageClassification""", """ConvNextModel""", """ConvNextPreTrainedModel""", """ConvNextBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : str = [ """TFConvNextForImageClassification""", """TFConvNextModel""", """TFConvNextPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _lowerCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
356
"""simple docstring""" import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowerCAmelCase : Optional[int] = logging.get_logger(__name__) _lowerCAmelCase : Optional[int] = { """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } _lowerCAmelCase : List[Any] = { """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } _lowerCAmelCase : int = { """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def SCREAMING_SNAKE_CASE__ ( snake_case : List[Any] )-> str: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = set() UpperCAmelCase__ : Optional[int] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ : Dict = char UpperCAmelCase__ : Tuple = set(snake_case ) return pairs class lowerCAmelCase__ ( __magic_name__ ): SCREAMING_SNAKE_CASE_ =VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ =PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[Any] , snake_case__ : int , snake_case__ : Union[str, Any] , snake_case__ : Tuple="<s>" , snake_case__ : List[Any]="</s>" , snake_case__ : Union[str, Any]="</s>" , snake_case__ : Union[str, Any]="<s>" , snake_case__ : Any="<unk>" , snake_case__ : int="<pad>" , snake_case__ : List[str]="<mask>" , **snake_case__ : Optional[int] , ): '''simple docstring''' super().__init__( bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , **snake_case__ , ) UpperCAmelCase__ : Dict = vocab_file UpperCAmelCase__ : Tuple = merges_file UpperCAmelCase__ : List[Any] = {} UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : Dict = 2 UpperCAmelCase__ : Dict = 3 self.add_from_file(snake_case__ ) UpperCAmelCase__ : Optional[Any] = {v: k for k, v in self.encoder.items()} with open(snake_case__ , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ : Tuple = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ : Optional[Any] = [tuple(merge.split()[:-1] ) for merge in merges] UpperCAmelCase__ : List[Any] = dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) UpperCAmelCase__ : Dict = {} def __a ( self : int , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ : str = [self.cls_token_id] UpperCAmelCase__ : Any = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __a ( self : List[str] , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None , snake_case__ : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) if token_ids_a is None: return [1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1, 1] + ([0] * len(snake_case__ )) + [1] def __a ( self : Union[str, Any] , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : Tuple = [self.sep_token_id] UpperCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __a ( self : List[str] ): '''simple docstring''' return len(self.encoder ) def __a ( self : Any ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def __a ( self : Dict , snake_case__ : Tuple ): '''simple docstring''' if token in self.cache: return self.cache[token] UpperCAmelCase__ : Optional[Any] = tuple(snake_case__ ) UpperCAmelCase__ : Optional[Any] = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) UpperCAmelCase__ : Any = get_pairs(snake_case__ ) if not pairs: return token while True: UpperCAmelCase__ : List[Any] = min(snake_case__ , key=lambda snake_case__ : self.bpe_ranks.get(snake_case__ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ : Tuple = bigram UpperCAmelCase__ : Optional[Any] = [] UpperCAmelCase__ : Tuple = 0 while i < len(snake_case__ ): try: UpperCAmelCase__ : Union[str, Any] = word.index(snake_case__ , snake_case__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ : Dict = j if word[i] == first and i < len(snake_case__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ : Dict = tuple(snake_case__ ) UpperCAmelCase__ : List[Any] = new_word if len(snake_case__ ) == 1: break else: UpperCAmelCase__ : Dict = get_pairs(snake_case__ ) UpperCAmelCase__ : List[Any] = "@@ ".join(snake_case__ ) UpperCAmelCase__ : Optional[int] = word[:-4] UpperCAmelCase__ : Union[str, Any] = word return word def __a ( self : List[Any] , snake_case__ : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : int = re.findall(R"\S+\n?" , snake_case__ ) for token in words: split_tokens.extend(list(self.bpe(snake_case__ ).split(" " ) ) ) return split_tokens def __a ( self : Dict , snake_case__ : List[str] ): '''simple docstring''' return self.encoder.get(snake_case__ , self.encoder.get(self.unk_token ) ) def __a ( self : List[Any] , snake_case__ : Any ): '''simple docstring''' return self.decoder.get(snake_case__ , self.unk_token ) def __a ( self : str , snake_case__ : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = " ".join(snake_case__ ).replace("@@ " , "" ).strip() return out_string def __a ( self : Any , snake_case__ : str , snake_case__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return UpperCAmelCase__ : Tuple = os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ : str = os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) if os.path.abspath(self.merges_file ) != os.path.abspath(snake_case__ ): copyfile(self.merges_file , snake_case__ ) return out_vocab_file, out_merge_file def __a ( self : List[Any] , snake_case__ : Union[str, Any] ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): try: with open(snake_case__ , "r" , encoding="utf-8" ) as fd: self.add_from_file(snake_case__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'Incorrect encoding detected in {f}, please rebuild the dataset' ) return UpperCAmelCase__ : Dict = f.readlines() for lineTmp in lines: UpperCAmelCase__ : Optional[int] = lineTmp.strip() UpperCAmelCase__ : Tuple = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) UpperCAmelCase__ : Any = line[:idx] UpperCAmelCase__ : str = len(self.encoder )
298
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class a__ ( unittest.TestCase ): def lowercase ( self : int ) -> Tuple: lowercase : List[Any] = 'laion/clap-htsat-unfused' lowercase : Tuple = tempfile.mkdtemp() def lowercase ( self : Union[str, Any], **lowerCAmelCase : List[str] ) -> Optional[int]: return RobertaTokenizer.from_pretrained(self.checkpoint, **_A ) def lowercase ( self : str, **lowerCAmelCase : Optional[int] ) -> List[Any]: return ClapFeatureExtractor.from_pretrained(self.checkpoint, **_A ) def lowercase ( self : str ) -> List[str]: shutil.rmtree(self.tmpdirname ) def lowercase ( self : Union[str, Any] ) -> Tuple: lowercase : Tuple = self.get_tokenizer() lowercase : Optional[Any] = self.get_feature_extractor() lowercase : Any = ClapProcessor(tokenizer=_A, feature_extractor=_A ) processor.save_pretrained(self.tmpdirname ) lowercase : Tuple = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer, _A ) self.assertEqual(processor.feature_extractor.to_json_string(), feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor, _A ) def lowercase ( self : Optional[int] ) -> int: lowercase : List[str] = ClapProcessor(tokenizer=self.get_tokenizer(), feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) lowercase : Optional[Any] = self.get_tokenizer(bos_token='(BOS)', eos_token='(EOS)' ) lowercase : str = self.get_feature_extractor(do_normalize=_A, padding_value=1.0 ) lowercase : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname, bos_token='(BOS)', eos_token='(EOS)', do_normalize=_A, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, _A ) self.assertEqual(processor.feature_extractor.to_json_string(), feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor, _A ) def lowercase ( self : Union[str, Any] ) -> Any: lowercase : Any = self.get_feature_extractor() lowercase : int = self.get_tokenizer() lowercase : Tuple = ClapProcessor(tokenizer=_A, feature_extractor=_A ) lowercase : List[Any] = floats_list((3, 1000) ) lowercase : Optional[Any] = feature_extractor(_A, return_tensors='np' ) lowercase : Dict = processor(audios=_A, return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) def lowercase ( self : List[Any] ) -> Optional[Any]: lowercase : Union[str, Any] = self.get_feature_extractor() lowercase : Tuple = self.get_tokenizer() lowercase : Any = ClapProcessor(tokenizer=_A, feature_extractor=_A ) lowercase : Tuple = 'This is a test string' lowercase : Any = processor(text=_A ) lowercase : List[Any] = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def lowercase ( self : int ) -> Optional[Any]: lowercase : Dict = self.get_feature_extractor() lowercase : List[Any] = self.get_tokenizer() lowercase : List[str] = ClapProcessor(tokenizer=_A, feature_extractor=_A ) lowercase : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowercase : Tuple = processor.batch_decode(_A ) lowercase : Any = tokenizer.batch_decode(_A ) self.assertListEqual(_A, _A ) def lowercase ( self : List[str] ) -> Any: lowercase : Optional[Any] = self.get_feature_extractor() lowercase : Tuple = self.get_tokenizer() lowercase : List[Any] = ClapProcessor(tokenizer=_A, feature_extractor=_A ) self.assertListEqual( processor.model_input_names[2:], feature_extractor.model_input_names, msg='`processor` and `feature_extractor` model input names do not match', )
255
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : int= logging.get_logger(__name__) _a : Optional[Any]= { "SCUT-DLVCLab/lilt-roberta-en-base": ( "https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json" ), } class UpperCamelCase ( lowercase ): UpperCAmelCase : List[Any] = """lilt""" def __init__(self : Dict , _A : Any=3_05_22 , _A : Union[str, Any]=7_68 , _A : Any=12 , _A : Tuple=12 , _A : Optional[int]=30_72 , _A : Tuple="gelu" , _A : str=0.1 , _A : List[Any]=0.1 , _A : Union[str, Any]=5_12 , _A : Any=2 , _A : Tuple=0.02 , _A : List[str]=1E-12 , _A : Optional[int]=0 , _A : Optional[Any]="absolute" , _A : Any=None , _A : List[Any]=4 , _A : Optional[int]=10_24 , **_A : Union[str, Any] , ) -> Tuple: super().__init__(pad_token_id=_A , **_A) __snake_case : Optional[int] = vocab_size __snake_case : List[Any] = hidden_size __snake_case : Any = num_hidden_layers __snake_case : Optional[int] = num_attention_heads __snake_case : Optional[int] = hidden_act __snake_case : List[str] = intermediate_size __snake_case : Union[str, Any] = hidden_dropout_prob __snake_case : Dict = attention_probs_dropout_prob __snake_case : List[Any] = max_position_embeddings __snake_case : Dict = type_vocab_size __snake_case : List[Any] = initializer_range __snake_case : Optional[Any] = layer_norm_eps __snake_case : Optional[int] = position_embedding_type __snake_case : Any = classifier_dropout __snake_case : Optional[int] = channel_shrink_ratio __snake_case : Tuple = max_ad_position_embeddings
172
0
"""simple docstring""" import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 A_ = get_tests_dir('''fixtures''') class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : Dict = mock.Mock() _snake_case : List[str] = 500 _snake_case : Dict = {} _snake_case : Optional[Any] = HTTPError _snake_case : List[Any] = {} # Download this model to make sure it's in the cache. _snake_case : int = ViTImageProcessor.from_pretrained("""hf-internal-testing/tiny-random-vit""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""", return_value=a_ ) as mock_head: _snake_case : Dict = ViTImageProcessor.from_pretrained("""hf-internal-testing/tiny-random-vit""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : List[Any] = ViTImageProcessor.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json""" ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' with self.assertRaises(a_ ): # config is in subfolder, the following should not work without specifying the subfolder _snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained("""hf-internal-testing/stable-diffusion-all-variants""" ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained( """hf-internal-testing/stable-diffusion-all-variants""", subfolder="""feature_extractor""" ) self.assertIsNotNone(a_ ) @is_staging_test class lowercase( unittest.TestCase ): '''simple docstring''' @classmethod def UpperCamelCase_ ( cls: int ): '''simple docstring''' _snake_case : List[Any] = TOKEN HfFolder.save_token(a_ ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' try: delete_repo(token=cls._token, repo_id="""test-image-processor""" ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id="""valid_org/test-image-processor-org""" ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id="""test-dynamic-image-processor""" ) except HTTPError: pass def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ViTImageProcessor.from_pretrained(a_ ) image_processor.push_to_hub("""test-image-processor""", use_auth_token=self._token ) _snake_case : int = ViTImageProcessor.from_pretrained(f"{USER}/test-image-processor" ) for k, v in image_processor.__dict__.items(): self.assertEqual(a_, getattr(a_, a_ ) ) # Reset repo delete_repo(token=self._token, repo_id="""test-image-processor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( a_, repo_id="""test-image-processor""", push_to_hub=a_, use_auth_token=self._token ) _snake_case : Any = ViTImageProcessor.from_pretrained(f"{USER}/test-image-processor" ) for k, v in image_processor.__dict__.items(): self.assertEqual(a_, getattr(a_, a_ ) ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : int = ViTImageProcessor.from_pretrained(a_ ) image_processor.push_to_hub("""valid_org/test-image-processor""", use_auth_token=self._token ) _snake_case : Tuple = ViTImageProcessor.from_pretrained("""valid_org/test-image-processor""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(a_, getattr(a_, a_ ) ) # Reset repo delete_repo(token=self._token, repo_id="""valid_org/test-image-processor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( a_, repo_id="""valid_org/test-image-processor-org""", push_to_hub=a_, use_auth_token=self._token ) _snake_case : Tuple = ViTImageProcessor.from_pretrained("""valid_org/test-image-processor-org""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(a_, getattr(a_, a_ ) ) def UpperCamelCase_ ( self: int ): '''simple docstring''' CustomImageProcessor.register_for_auto_class() _snake_case : int = CustomImageProcessor.from_pretrained(a_ ) image_processor.push_to_hub("""test-dynamic-image-processor""", use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map, {"""AutoImageProcessor""": """custom_image_processing.CustomImageProcessor"""}, ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained( f"{USER}/test-dynamic-image-processor", trust_remote_code=a_ ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__, """CustomImageProcessor""" )
351
"""simple docstring""" from collections.abc import Generator from math import sin def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" if len(snake_case__ ) != 32: raise ValueError("""Input must be of length 32""" ) _snake_case : Optional[int] = B"""""" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) _snake_case : Optional[Any] = format(snake_case__ , """08x""" )[-8:] _snake_case : Optional[Any] = B"""""" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("""utf-8""" ) return little_endian_hex def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" _snake_case : Union[str, Any] = B"""""" for char in message: bit_string += format(snake_case__ , """08b""" ).encode("""utf-8""" ) _snake_case : List[Any] = format(len(snake_case__ ) , """064b""" ).encode("""utf-8""" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__ ) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" if len(snake_case__ ) % 5_12 != 0: raise ValueError("""Input must have length that's a multiple of 512""" ) for pos in range(0 , len(snake_case__ ) , 5_12 ): _snake_case : List[str] = bit_string[pos : pos + 5_12] _snake_case : List[str] = [] for i in range(0 , 5_12 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) _snake_case : Optional[int] = format(snake_case__ , """032b""" ) _snake_case : str = """""" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2 ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : int ): """simple docstring""" return (a + b) % 2**32 def UpperCAmelCase__ (snake_case__ : int , snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) if shift < 0: raise ValueError("""Shift must be non-negative""" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" _snake_case : Any = preprocess(snake_case__ ) _snake_case : Optional[Any] = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states _snake_case : Union[str, Any] = 0x6745_2301 _snake_case : List[Any] = 0xEFCD_AB89 _snake_case : Optional[Any] = 0x98BA_DCFE _snake_case : Optional[int] = 0x1032_5476 _snake_case : Tuple = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__ ): _snake_case : Tuple = aa _snake_case : str = ba _snake_case : int = ca _snake_case : Dict = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f _snake_case : int = d ^ (b & (c ^ d)) _snake_case : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f _snake_case : Tuple = c ^ (d & (b ^ c)) _snake_case : int = (5 * i + 1) % 16 elif i <= 47: _snake_case : List[Any] = b ^ c ^ d _snake_case : Union[str, Any] = (3 * i + 5) % 16 else: _snake_case : Tuple = c ^ (b | not_aa(snake_case__ )) _snake_case : Optional[int] = (7 * i) % 16 _snake_case : Dict = (f + a + added_consts[i] + block_words[g]) % 2**32 _snake_case : List[str] = d _snake_case : List[Any] = c _snake_case : str = b _snake_case : List[str] = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i] ) ) # Add hashed chunk to running total _snake_case : Union[str, Any] = sum_aa(snake_case__ , snake_case__ ) _snake_case : str = sum_aa(snake_case__ , snake_case__ ) _snake_case : Any = sum_aa(snake_case__ , snake_case__ ) _snake_case : List[str] = sum_aa(snake_case__ , snake_case__ ) _snake_case : Any = reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) return digest if __name__ == "__main__": import doctest doctest.testmod()
132
0
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class UpperCAmelCase__ ( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" a = StableUnCLIPPipeline a = TEXT_TO_IMAGE_PARAMS a = TEXT_TO_IMAGE_BATCH_PARAMS a = TEXT_TO_IMAGE_IMAGE_PARAMS a = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false a = False def lowercase_ ( self : Dict ) -> List[str]: SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = embedder_hidden_size # prior components torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowerCamelCase , projection_dim=__lowerCamelCase , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=__lowerCamelCase , num_layers=1 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=__lowerCamelCase , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = StableUnCLIPImageNormalizer(embedding_dim=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowerCamelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowerCamelCase , layers_per_block=1 , upcast_attention=__lowerCamelCase , use_linear_projection=__lowerCamelCase , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.00085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowerCamelCase , steps_offset=1 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL() SCREAMING_SNAKE_CASE__ = { # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def lowercase_ ( self : List[str] , __lowerCamelCase : List[str] , __lowerCamelCase : Any=0 ) -> Any: if str(__lowerCamelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(__lowerCamelCase ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def lowercase_ ( self : List[Any] ) -> List[str]: SCREAMING_SNAKE_CASE__ = torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=__lowerCamelCase ) def lowercase_ ( self : Optional[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowerCamelCase ) @slow @require_torch_gpu class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self : str ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self : Optional[int] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) SCREAMING_SNAKE_CASE__ = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE__ = torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ = pipe('''anime turle''' , generator=__lowerCamelCase , output_type='''np''' ) SCREAMING_SNAKE_CASE__ = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowerCamelCase , __lowerCamelCase ) def lowercase_ ( self : Dict ) -> List[str]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() SCREAMING_SNAKE_CASE__ = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE__ = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE__ = pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) SCREAMING_SNAKE_CASE__ = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
218
import warnings from .generation import TFGenerationMixin class UpperCAmelCase__ ( A__ ): """simple docstring""" warnings.warn( "Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will " "be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead." , A__ , )
218
1
"""simple docstring""" import argparse import torch from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert from transformers.utils import logging logging.set_verbosity_info() def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : List[str] , _lowercase : Dict ) ->int: '''simple docstring''' a : Dict = BertConfig.from_json_file(_lowercase ) print(F"""Building PyTorch model from configuration: {config}""" ) a : Optional[Any] = BertForPreTraining(_lowercase ) # Load weights from tf checkpoint load_tf_weights_in_bert(_lowercase , _lowercase , _lowercase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowercase ) if __name__ == "__main__": a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--bert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) a : Optional[int] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
105
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = os.path.join(args.tf_model_dir , "parameters.json" ) a : str = json.loads(open(_lowercase ).read() ) if not params: raise ValueError( F"""It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.""" ) if not args.output.endswith(".pt" ): a : str = args.output + ".pt" a : Dict = OrderedDict() with tf.device("/CPU:0" ): a : Optional[int] = tf.train.load_checkpoint(args.tf_model_dir ) a : Optional[Any] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): a : Dict = reader.get_tensor(_lowercase ).astype(np.floataa ) if key_name.endswith("/adam_m" ) or key_name.endswith("/adam_v" ): continue if key_name.startswith("pasts/" ): if key_name.startswith("pasts/mlp" ): a : Union[str, Any] = int(key_name[9] ) elif key_name.startswith("pasts/out" ): a : Optional[int] = 8 a : Dict = "model.sqout.%d.weight" % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time a : int = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : Dict = torch.tensor(_lowercase ) elif key_name.startswith("model/moe" ): a : List[str] = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/switch_gating/kernel" ): a : str = "model.blocks.%d.feed_forward.mlp.router.classifier.weight" % player a : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : List[str] = torch.tensor(_lowercase ) elif key_name.endswith("/softmlp/kernel" ): a : Optional[int] = "model.blocks.%d.feed_forward.soft_bypass_mlp.weight" % player a : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : List[Any] = torch.tensor(_lowercase ) elif key_name.endswith("/wo/kernel" ) or key_name.endswith("/wi/kernel" ): a : Any = key_name[-9:-7] for i in range(16 ): a : List[Any] = "model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight" % (player, i, nlayer) a : str = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided a : Dict = torch.tensor(_lowercase ) elif key_name.startswith("model/mlp" ): a : Union[str, Any] = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/p1/kernel" ): a : str = "model.blocks.%d.feed_forward.mlp.wi.weight" % player a : Dict = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : Optional[int] = torch.tensor(_lowercase ) elif key_name.endswith("/p1/bias" ): a : str = "model.blocks.%d.feed_forward.mlp.wi.bias" % player a : List[Any] = vnp.copy() # same because it is one dimensional a : Tuple = torch.tensor(_lowercase ) elif key_name.endswith("/p2/kernel" ): a : Union[str, Any] = "model.blocks.%d.feed_forward.mlp.wo.weight" % player a : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : List[Any] = torch.tensor(_lowercase ) elif key_name.endswith("/p2/bias" ): a : Dict = "model.blocks.%d.feed_forward.mlp.wo.bias" % player a : List[str] = vnp.copy() # same because it is one dimensional a : str = torch.tensor(_lowercase ) elif key_name.startswith("model/ln" ): a : List[str] = int(key_name[8:].split("/" )[0] ) if key_name.endswith("/b" ): a : Optional[Any] = "model.blocks.%d.feed_forward.norm.bias" % player a : Tuple = vnp.copy() # same because it is one dimensional a : int = torch.tensor(_lowercase ) elif key_name.endswith("/g" ): a : Optional[Any] = "model.blocks.%d.feed_forward.norm.weight" % player a : List[str] = vnp.copy() # same because it is one dimensional a : Tuple = torch.tensor(_lowercase ) elif key_name.startswith("model/att" ): a : Optional[Any] = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/qkv/kernel" ): a : Union[str, Any] = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum a : List[str] = state[:, 0, :, :] a : Dict = state[:, 1, :, :] a : Union[str, Any] = state[:, 2, :, :] a : str = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix a : List[str] = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix a : Dict = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix a : List[Any] = "model.blocks.%d.self_attn.self_attn.q_proj.weight" % player a : Union[str, Any] = torch.tensor(_lowercase ) a : Any = "model.blocks.%d.self_attn.self_attn.k_proj.weight" % player a : List[str] = torch.tensor(_lowercase ) a : Optional[Any] = "model.blocks.%d.self_attn.self_attn.v_proj.weight" % player a : Optional[Any] = torch.tensor(_lowercase ) elif key_name.endswith("/o/kernel" ): a : Any = "model.blocks.%d.self_attn.self_attn.out_proj.weight" % player a : Optional[int] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix a : Tuple = torch.tensor(_lowercase ) elif key_name.startswith("model/an" ): a : List[str] = int(key_name[8:].split("/" )[0] ) if key_name.endswith("/b" ): a : Optional[int] = "model.blocks.%d.self_attn.norm.bias" % player a : Union[str, Any] = vnp.copy() # same because it is one dimensional a : List[Any] = torch.tensor(_lowercase ) elif key_name.endswith("/g" ): a : Any = "model.blocks.%d.self_attn.norm.weight" % player a : str = vnp.copy() # same because it is one dimensional a : Any = torch.tensor(_lowercase ) elif ( key_name.startswith("model/wte" ) or key_name.startswith("model/wpe" ) or key_name.startswith("model/ete" ) ): a : Optional[int] = {"wte": "embed_tokens", "wpe": "position_embeddings", "ete": "extra_position_embeddings"}[ key_name[-3:] ] a : Tuple = "model.%s.weight" % nlayer a : Any = vnp.copy() # same in embedded a : Tuple = torch.tensor(_lowercase ) if key_name.startswith("model/wte" ): a : Optional[int] = "lm_head.weight" a : Optional[int] = vnp.copy() # same in embedded a : Optional[int] = torch.tensor(_lowercase ) elif key_name.startswith("model/wob" ): a : Optional[int] = "final_logits_bias" a : Optional[Any] = vnp.copy() # same in embedded a : Optional[int] = state.reshape((1, -1) ) a : List[Any] = torch.tensor(_lowercase ) elif key_name == "model/dense/kernel": a : Optional[int] = "model.last_project.weight" a : Union[str, Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix a : List[Any] = torch.tensor(_lowercase ) elif key_name == "model/dense_1/bias": a : Dict = "model.last_project.bias" a : Optional[Any] = vnp.copy() # same because it is one dimensional a : Any = torch.tensor(_lowercase ) torch.save(_lowercase , args.output ) if __name__ == "__main__": a : Optional[int] = argparse.ArgumentParser( description='''model converter.''', formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument('''--tf_model_dir''', metavar='''PATH''', type=str, required=True, help='''import model''') parser.add_argument('''--output''', metavar='''PATH''', type=str, required=True, help='''output model''') a : Tuple = parser.parse_args() convert_tf_gptsan_to_pt(args)
105
1
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP __A = False try: __A = _is_package_available("google.colab") except ModuleNotFoundError: pass @input.register class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : str , UpperCAmelCase_ : str = None , UpperCAmelCase_ : list = []) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Union[str, Any] =0 lowerCamelCase__: Dict =choices lowerCamelCase__: Union[str, Any] =prompt if sys.platform == "win32": lowerCamelCase__: Optional[Any] ="*" else: lowerCamelCase__: int ="➔ " def SCREAMING_SNAKE_CASE_ (self : str , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str = "") ->Any: '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 32 , UpperCAmelCase_) else: forceWrite(self.choices[index] , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Dict , UpperCAmelCase_ : int) ->Optional[int]: '''simple docstring''' if index == self.position: forceWrite(F""" {self.arrow_char} """) self.write_choice(UpperCAmelCase_) else: forceWrite(F""" {self.choices[index]}""") reset_cursor() def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : Direction , UpperCAmelCase_ : int = 1) ->Any: '''simple docstring''' lowerCamelCase__: Optional[int] =self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(UpperCAmelCase_) move_cursor(UpperCAmelCase_ , direction.name) self.print_choice(self.position) @input.mark(KEYMAP["up"]) def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->str: '''simple docstring''' self.move_direction(Direction.UP) @input.mark(KEYMAP["down"]) def SCREAMING_SNAKE_CASE_ (self : str) ->str: '''simple docstring''' self.move_direction(Direction.DOWN) @input.mark(KEYMAP["newline"]) def SCREAMING_SNAKE_CASE_ (self : int) ->int: '''simple docstring''' move_cursor(len(self.choices) - self.position , "DOWN") return self.position @input.mark(KEYMAP["interrupt"]) def SCREAMING_SNAKE_CASE_ (self : Any) ->Tuple: '''simple docstring''' move_cursor(len(self.choices) - self.position , "DOWN") raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(UpperCAmelCase_)] for number in range(10)]) def SCREAMING_SNAKE_CASE_ (self : str) ->str: '''simple docstring''' lowerCamelCase__: str =int(chr(self.current_selection)) lowerCamelCase__: Optional[int] =index - self.position if index == self.position: return if index < len(self.choices): if self.position > index: self.move_direction(Direction.UP , -movement) elif self.position < index: self.move_direction(Direction.DOWN , UpperCAmelCase_) else: return else: return def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : int = 0) ->Union[str, Any]: '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , "\n") if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n") else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n") lowerCamelCase__: Dict =default_choice for i in range(len(self.choices)): self.print_choice(UpperCAmelCase_) forceWrite("\n") move_cursor(len(self.choices) - self.position , "UP") with cursor.hide(): while True: if in_colab: try: lowerCamelCase__: Optional[Any] =int(builtins.input()) except ValueError: lowerCamelCase__: Optional[Any] =default_choice else: lowerCamelCase__: List[Any] =self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices) + 1): move_cursor(1 , "UP") clear_line() self.write_choice(UpperCAmelCase_ , "\n") return choice
273
def lowerCAmelCase_ ( __a , __a ) -> Tuple: """simple docstring""" assert x is not None assert y is not None lowerCamelCase__: Any =len(__a ) lowerCamelCase__: int =len(__a ) # declaring the array for storing the dp values lowerCamelCase__: List[Any] =[[0] * (n + 1) for _ in range(m + 1 )] # noqa: E741 for i in range(1 , m + 1 ): for j in range(1 , n + 1 ): lowerCamelCase__: str =1 if x[i - 1] == y[j - 1] else 0 lowerCamelCase__: str =max(l[i - 1][j] , l[i][j - 1] , l[i - 1][j - 1] + match ) lowerCamelCase__: Any ="" lowerCamelCase__ , lowerCamelCase__: str =m, n while i > 0 and j > 0: lowerCamelCase__: Union[str, Any] =1 if x[i - 1] == y[j - 1] else 0 if l[i][j] == l[i - 1][j - 1] + match: if match == 1: lowerCamelCase__: Any =x[i - 1] + seq i -= 1 j -= 1 elif l[i][j] == l[i - 1][j]: i -= 1 else: j -= 1 return l[m][n], seq if __name__ == "__main__": __A = "AGGTAB" __A = "GXTXAYB" __A = 4 __A = "GTAB" __A , __A = longest_common_subsequence(a, b) print("len =", ln, ", sub-sequence =", subseq) import doctest doctest.testmod()
273
1
'''simple docstring''' import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets UpperCAmelCase_ = datasets.logging.get_logger(__name__) UpperCAmelCase_ = '\\n@inproceedings{bleurt,\n title={BLEURT: Learning Robust Metrics for Text Generation},\n author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh},\n booktitle={ACL},\n year={2020},\n url={https://arxiv.org/abs/2004.04696}\n}\n' UpperCAmelCase_ = '\\nBLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018)\nand then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune\nit for your specific application (the latter is expected to perform better).\n\nSee the project\'s README at https://github.com/google-research/bleurt#readme for more information.\n' UpperCAmelCase_ = '\nBLEURT score.\n\nArgs:\n `predictions` (list of str): prediction/candidate sentences\n `references` (list of str): reference sentences\n `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None.\n\nReturns:\n \'scores\': List of scores.\nExamples:\n\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> bleurt = datasets.load_metric("bleurt")\n >>> results = bleurt.compute(predictions=predictions, references=references)\n >>> print([round(v, 2) for v in results["scores"]])\n [1.03, 1.04]\n' UpperCAmelCase_ = { 'bleurt-tiny-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip', 'bleurt-tiny-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip', 'bleurt-base-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip', 'bleurt-base-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip', 'bleurt-large-128': 'https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip', 'bleurt-large-512': 'https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip', 'BLEURT-20-D3': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip', 'BLEURT-20-D6': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip', 'BLEURT-20-D12': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip', 'BLEURT-20': 'https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/google-research/bleurt""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/google-research/bleurt"""] , reference_urls=["""https://github.com/google-research/bleurt""", """https://arxiv.org/abs/2004.04696"""] , ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , _UpperCAmelCase : Optional[int] ): """simple docstring""" if self.config_name == "default": logger.warning( """Using default BLEURT-Base checkpoint for sequence maximum length 128. """ """You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512').""" ) UpperCAmelCase__ = """bleurt-base-128""" if self.config_name.lower() in CHECKPOINT_URLS: UpperCAmelCase__ = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: UpperCAmelCase__ = self.config_name.upper() else: raise KeyError( f'''{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}''' ) # download the model checkpoint specified by self.config_name and set up the scorer UpperCAmelCase__ = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) UpperCAmelCase__ = score.BleurtScorer(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any] ): """simple docstring""" UpperCAmelCase__ = self.scorer.score(references=_UpperCAmelCase , candidates=_UpperCAmelCase ) return {"scores": scores}
346
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' @register_to_config def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : float , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : str , _UpperCAmelCase : bool = False , ): """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = False UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) UpperCAmelCase__ = TaConfig( vocab_size=_UpperCAmelCase , d_model=_UpperCAmelCase , num_heads=_UpperCAmelCase , d_kv=_UpperCAmelCase , d_ff=_UpperCAmelCase , dropout_rate=_UpperCAmelCase , feed_forward_proj=_UpperCAmelCase , is_decoder=_UpperCAmelCase , is_encoder_decoder=_UpperCAmelCase , ) UpperCAmelCase__ = nn.ModuleList() for lyr_num in range(_UpperCAmelCase ): UpperCAmelCase__ = TaBlock(_UpperCAmelCase ) self.encoders.append(_UpperCAmelCase ) UpperCAmelCase__ = TaLayerNorm(_UpperCAmelCase ) UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Dict , _UpperCAmelCase : int , _UpperCAmelCase : str ): """simple docstring""" UpperCAmelCase__ = self.token_embedder(_UpperCAmelCase ) UpperCAmelCase__ = encoder_input_tokens.shape[1] UpperCAmelCase__ = torch.arange(_UpperCAmelCase , device=encoder_input_tokens.device ) x += self.position_encoding(_UpperCAmelCase ) UpperCAmelCase__ = self.dropout_pre(_UpperCAmelCase ) # inverted the attention mask UpperCAmelCase__ = encoder_input_tokens.size() UpperCAmelCase__ = self.get_extended_attention_mask(_UpperCAmelCase , _UpperCAmelCase ) for lyr in self.encoders: UpperCAmelCase__ = lyr(_UpperCAmelCase , _UpperCAmelCase )[0] UpperCAmelCase__ = self.layer_norm(_UpperCAmelCase ) return self.dropout_post(_UpperCAmelCase ), encoder_inputs_mask
346
1
"""simple docstring""" import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( enum.Enum ): lowercase__ = 0 lowercase__ = 1 @add_end_docstrings(__UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "generated" def __init__( self : Union[str, Any] , *lowerCAmelCase_ : int , **lowerCAmelCase_ : str): """simple docstring""" super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == """tf""" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def _UpperCAmelCase ( self : List[Any] , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : List[Any]=None , lowerCAmelCase_ : str=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : List[Any]=None , **lowerCAmelCase_ : Tuple , ): """simple docstring""" lowercase_ = {} if truncation is not None: lowercase_ = truncation lowercase_ = generate_kwargs lowercase_ = {} if return_tensors is not None and return_type is None: lowercase_ = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: lowercase_ = return_type if clean_up_tokenization_spaces is not None: lowercase_ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase_ = self.tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_) if len(lowerCAmelCase_) > 1: warnings.warn( """Stopping on a multiple token sequence is not yet supported on transformers. The first token of""" """ the stop sequence will be used as the stop sequence string in the interim.""") lowercase_ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _UpperCAmelCase ( self : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int): """simple docstring""" return True def _UpperCAmelCase ( self : Optional[Any] , *lowerCAmelCase_ : int , lowerCAmelCase_ : Dict): """simple docstring""" lowercase_ = self.model.config.prefix if self.model.config.prefix is not None else """""" if isinstance(args[0] , lowerCAmelCase_): if self.tokenizer.pad_token_id is None: raise ValueError("""Please make sure that the tokenizer has a pad_token_id when using a batch input""") lowercase_ = ([prefix + arg for arg in args[0]],) lowercase_ = True elif isinstance(args[0] , lowerCAmelCase_): lowercase_ = (prefix + args[0],) lowercase_ = False else: raise ValueError( F''' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`''') lowercase_ = self.tokenizer(*lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Union[str, Any] , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Any): """simple docstring""" lowercase_ = super().__call__(*lowerCAmelCase_ , **lowerCAmelCase_) if ( isinstance(args[0] , lowerCAmelCase_) and all(isinstance(lowerCAmelCase_ , lowerCAmelCase_) for el in args[0]) and all(len(lowerCAmelCase_) == 1 for res in result) ): return [res[0] for res in result] return result def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str]=TruncationStrategy.DO_NOT_TRUNCATE , **lowerCAmelCase_ : Dict): """simple docstring""" lowercase_ = self._parse_and_tokenize(lowerCAmelCase_ , truncation=lowerCAmelCase_ , **lowerCAmelCase_) return inputs def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : List[str] , **lowerCAmelCase_ : Optional[Any]): """simple docstring""" if self.framework == "pt": lowercase_ , lowercase_ = model_inputs["""input_ids"""].shape elif self.framework == "tf": lowercase_ , lowercase_ = tf.shape(model_inputs["""input_ids"""]).numpy() lowercase_ = generate_kwargs.get("""min_length""" , self.model.config.min_length) lowercase_ = generate_kwargs.get("""max_length""" , self.model.config.max_length) self.check_inputs(lowerCAmelCase_ , generate_kwargs["""min_length"""] , generate_kwargs["""max_length"""]) lowercase_ = self.model.generate(**lowerCAmelCase_ , **lowerCAmelCase_) lowercase_ = output_ids.shape[0] if self.framework == "pt": lowercase_ = output_ids.reshape(lowerCAmelCase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": lowercase_ = tf.reshape(lowerCAmelCase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : str=ReturnType.TEXT , lowerCAmelCase_ : int=False): """simple docstring""" lowercase_ = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: lowercase_ = {F'''{self.return_name}_token_ids''': output_ids} elif return_type == ReturnType.TEXT: lowercase_ = { F'''{self.return_name}_text''': self.tokenizer.decode( lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ , ) } records.append(lowerCAmelCase_) return records @add_end_docstrings(__UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "summary" def __call__( self : Union[str, Any] , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : Any): """simple docstring""" return super().__call__(*lowerCAmelCase_ , **lowerCAmelCase_) def _UpperCAmelCase ( self : Tuple , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int): """simple docstring""" if max_length < min_length: logger.warning(F'''Your min_length={min_length} must be inferior than your max_length={max_length}.''') if input_length < max_length: logger.warning( F'''Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ''' """a summarization task, where outputs shorter than the input are typically wanted, you might """ F'''consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})''') @add_end_docstrings(__UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "translation" def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int): """simple docstring""" if input_length > 0.9 * max_length: logger.warning( F'''Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ''' """increasing your max_length manually, e.g. translator('...', max_length=400)""") return True def _UpperCAmelCase ( self : Union[str, Any] , *lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Any=TruncationStrategy.DO_NOT_TRUNCATE , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : Tuple=None): """simple docstring""" if getattr(self.tokenizer , """_build_translation_inputs""" , lowerCAmelCase_): return self.tokenizer._build_translation_inputs( *lowerCAmelCase_ , return_tensors=self.framework , truncation=lowerCAmelCase_ , src_lang=lowerCAmelCase_ , tgt_lang=lowerCAmelCase_) else: return super()._parse_and_tokenize(*lowerCAmelCase_ , truncation=lowerCAmelCase_) def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : int=None , **lowerCAmelCase_ : Tuple): """simple docstring""" lowercase_ , lowercase_ , lowercase_ = super()._sanitize_parameters(**lowerCAmelCase_) if src_lang is not None: lowercase_ = src_lang if tgt_lang is not None: lowercase_ = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. lowercase_ = kwargs.get("""task""" , self.task) lowercase_ = task.split("""_""") if task and len(lowerCAmelCase_) == 4: # translation, XX, to YY lowercase_ = items[1] lowercase_ = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : int , *lowerCAmelCase_ : int , **lowerCAmelCase_ : Union[str, Any]): """simple docstring""" return super().__call__(*lowerCAmelCase_ , **lowerCAmelCase_)
350
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html UpperCAmelCase : Optional[Any] = "platform" import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class SCREAMING_SNAKE_CASE__ : lowercase__ = PegasusConfig lowercase__ = {} lowercase__ = "gelu" def __init__( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=1_3 , lowerCAmelCase_ : Any=7 , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : str=9_9 , lowerCAmelCase_ : Tuple=3_2 , lowerCAmelCase_ : Dict=5 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Dict=3_7 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Optional[int]=2_0 , lowerCAmelCase_ : Tuple=2 , lowerCAmelCase_ : List[str]=1 , lowerCAmelCase_ : Optional[Any]=0 , ): """simple docstring""" lowercase_ = parent lowercase_ = batch_size lowercase_ = seq_length lowercase_ = is_training lowercase_ = use_labels lowercase_ = vocab_size lowercase_ = hidden_size lowercase_ = num_hidden_layers lowercase_ = num_attention_heads lowercase_ = intermediate_size lowercase_ = hidden_dropout_prob lowercase_ = attention_probs_dropout_prob lowercase_ = max_position_embeddings lowercase_ = eos_token_id lowercase_ = pad_token_id lowercase_ = bos_token_id def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" lowercase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size).clip(3 , self.vocab_size) lowercase_ = np.expand_dims(np.array([self.eos_token_id] * self.batch_size) , 1) lowercase_ = np.concatenate([input_ids, eos_tensor] , axis=1) lowercase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowercase_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowercase_ = prepare_pegasus_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) return config, inputs_dict def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Optional[Any]): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""") lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1])), ] , axis=-1 , ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowercase_ = np.not_equal(__lowerCAmelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: lowercase_ = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , unittest.TestCase ): lowercase__ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) lowercase__ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ = FlaxPegasusModelTester(self) lowercase_ = ConfigTester(self , config_class=lowerCAmelCase_) def _UpperCAmelCase ( self : Any): """simple docstring""" self.config_tester.run_common_tests() def _UpperCAmelCase ( self : List[str]): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model_class(lowerCAmelCase_) @jax.jit def encode_jitted(lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[int]=None , **lowerCAmelCase_ : Optional[int]): return model.encode(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_) with self.subTest("""JIT Enabled"""): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = model_class(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""]) lowercase_ = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict): return model.decode( decoder_input_ids=lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , encoder_outputs=lowerCAmelCase_ , ) with self.subTest("""JIT Enabled"""): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) @slow def _UpperCAmelCase ( self : Tuple): """simple docstring""" for model_class_name in self.all_model_classes: lowercase_ = model_class_name.from_pretrained("""google/pegasus-large""" , from_pt=lowerCAmelCase_) lowercase_ = np.ones((1, 1)) lowercase_ = model(lowerCAmelCase_) self.assertIsNotNone(lowerCAmelCase_) @slow def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""") lowercase_ = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""") lowercase_ = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] lowercase_ = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] lowercase_ = tokenizer(lowerCAmelCase_ , return_tensors="""np""" , truncation=lowerCAmelCase_ , max_length=5_1_2 , padding=lowerCAmelCase_) lowercase_ = model.generate(**lowerCAmelCase_ , num_beams=2).sequences lowercase_ = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_) assert tgt_text == decoded
313
0